var/home/core/zuul-output/0000755000175000017500000000000015070677506014542 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070724642015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006533535515070724632017722 0ustar rootrootOct 06 08:39:07 crc systemd[1]: Starting Kubernetes Kubelet... Oct 06 08:39:08 crc restorecon[4733]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:08 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 06 08:39:09 crc restorecon[4733]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 06 08:39:09 crc kubenswrapper[4989]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.672436 4989 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682637 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682697 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682704 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682710 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682717 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682724 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682729 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682734 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682740 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682746 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682753 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682759 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682764 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682769 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682774 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682779 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682784 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682789 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682794 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682799 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682804 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682809 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682814 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682821 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682826 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682842 4989 feature_gate.go:330] unrecognized feature gate: Example Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682848 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682853 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682857 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682862 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682867 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682872 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682877 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682882 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682887 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682891 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682898 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682906 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682914 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682922 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682927 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682934 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682939 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682943 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682949 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682953 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682961 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682967 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682973 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682978 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682983 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682988 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682993 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.682998 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683003 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683008 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683013 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683018 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683023 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683028 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683034 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683041 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683046 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683051 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683056 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683061 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683066 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683071 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683075 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683080 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.683087 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683200 4989 flags.go:64] FLAG: --address="0.0.0.0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683212 4989 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683222 4989 flags.go:64] FLAG: --anonymous-auth="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683230 4989 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683237 4989 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683243 4989 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683250 4989 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683258 4989 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683264 4989 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683270 4989 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683277 4989 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683284 4989 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683290 4989 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683296 4989 flags.go:64] FLAG: --cgroup-root="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683302 4989 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683308 4989 flags.go:64] FLAG: --client-ca-file="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683314 4989 flags.go:64] FLAG: --cloud-config="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683320 4989 flags.go:64] FLAG: --cloud-provider="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683325 4989 flags.go:64] FLAG: --cluster-dns="[]" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683333 4989 flags.go:64] FLAG: --cluster-domain="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683338 4989 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683344 4989 flags.go:64] FLAG: --config-dir="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683350 4989 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683356 4989 flags.go:64] FLAG: --container-log-max-files="5" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683364 4989 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683369 4989 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683376 4989 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683384 4989 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683391 4989 flags.go:64] FLAG: --contention-profiling="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683398 4989 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683405 4989 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683412 4989 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683419 4989 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683428 4989 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683436 4989 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683441 4989 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683447 4989 flags.go:64] FLAG: --enable-load-reader="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683452 4989 flags.go:64] FLAG: --enable-server="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683458 4989 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683465 4989 flags.go:64] FLAG: --event-burst="100" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683471 4989 flags.go:64] FLAG: --event-qps="50" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683476 4989 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683482 4989 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683487 4989 flags.go:64] FLAG: --eviction-hard="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683494 4989 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683500 4989 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683507 4989 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683513 4989 flags.go:64] FLAG: --eviction-soft="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683519 4989 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683525 4989 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683530 4989 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683535 4989 flags.go:64] FLAG: --experimental-mounter-path="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683541 4989 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683546 4989 flags.go:64] FLAG: --fail-swap-on="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683552 4989 flags.go:64] FLAG: --feature-gates="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683559 4989 flags.go:64] FLAG: --file-check-frequency="20s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683564 4989 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683570 4989 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683576 4989 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683582 4989 flags.go:64] FLAG: --healthz-port="10248" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683588 4989 flags.go:64] FLAG: --help="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683593 4989 flags.go:64] FLAG: --hostname-override="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683599 4989 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683604 4989 flags.go:64] FLAG: --http-check-frequency="20s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683610 4989 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683616 4989 flags.go:64] FLAG: --image-credential-provider-config="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683622 4989 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683627 4989 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683633 4989 flags.go:64] FLAG: --image-service-endpoint="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683638 4989 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683644 4989 flags.go:64] FLAG: --kube-api-burst="100" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683670 4989 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683676 4989 flags.go:64] FLAG: --kube-api-qps="50" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683682 4989 flags.go:64] FLAG: --kube-reserved="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683688 4989 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683720 4989 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683728 4989 flags.go:64] FLAG: --kubelet-cgroups="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683735 4989 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683742 4989 flags.go:64] FLAG: --lock-file="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683748 4989 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683754 4989 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683760 4989 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683775 4989 flags.go:64] FLAG: --log-json-split-stream="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683782 4989 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683788 4989 flags.go:64] FLAG: --log-text-split-stream="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683793 4989 flags.go:64] FLAG: --logging-format="text" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683799 4989 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683805 4989 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683811 4989 flags.go:64] FLAG: --manifest-url="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683816 4989 flags.go:64] FLAG: --manifest-url-header="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683824 4989 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683829 4989 flags.go:64] FLAG: --max-open-files="1000000" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683836 4989 flags.go:64] FLAG: --max-pods="110" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683842 4989 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683848 4989 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683854 4989 flags.go:64] FLAG: --memory-manager-policy="None" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683860 4989 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683865 4989 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683879 4989 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683887 4989 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683904 4989 flags.go:64] FLAG: --node-status-max-images="50" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683910 4989 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683915 4989 flags.go:64] FLAG: --oom-score-adj="-999" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683921 4989 flags.go:64] FLAG: --pod-cidr="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683927 4989 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683936 4989 flags.go:64] FLAG: --pod-manifest-path="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683942 4989 flags.go:64] FLAG: --pod-max-pids="-1" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683948 4989 flags.go:64] FLAG: --pods-per-core="0" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683954 4989 flags.go:64] FLAG: --port="10250" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683961 4989 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683968 4989 flags.go:64] FLAG: --provider-id="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683976 4989 flags.go:64] FLAG: --qos-reserved="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683983 4989 flags.go:64] FLAG: --read-only-port="10255" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683990 4989 flags.go:64] FLAG: --register-node="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.683997 4989 flags.go:64] FLAG: --register-schedulable="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684003 4989 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684014 4989 flags.go:64] FLAG: --registry-burst="10" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684019 4989 flags.go:64] FLAG: --registry-qps="5" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684025 4989 flags.go:64] FLAG: --reserved-cpus="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684031 4989 flags.go:64] FLAG: --reserved-memory="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684039 4989 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684075 4989 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684081 4989 flags.go:64] FLAG: --rotate-certificates="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684087 4989 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684092 4989 flags.go:64] FLAG: --runonce="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684099 4989 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684105 4989 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684111 4989 flags.go:64] FLAG: --seccomp-default="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684117 4989 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684122 4989 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684129 4989 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684138 4989 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684144 4989 flags.go:64] FLAG: --storage-driver-password="root" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684149 4989 flags.go:64] FLAG: --storage-driver-secure="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684155 4989 flags.go:64] FLAG: --storage-driver-table="stats" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684160 4989 flags.go:64] FLAG: --storage-driver-user="root" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684166 4989 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684173 4989 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684178 4989 flags.go:64] FLAG: --system-cgroups="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684184 4989 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684194 4989 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684200 4989 flags.go:64] FLAG: --tls-cert-file="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684206 4989 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684214 4989 flags.go:64] FLAG: --tls-min-version="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684219 4989 flags.go:64] FLAG: --tls-private-key-file="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684225 4989 flags.go:64] FLAG: --topology-manager-policy="none" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684230 4989 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684236 4989 flags.go:64] FLAG: --topology-manager-scope="container" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684242 4989 flags.go:64] FLAG: --v="2" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684250 4989 flags.go:64] FLAG: --version="false" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684257 4989 flags.go:64] FLAG: --vmodule="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684264 4989 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684270 4989 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684421 4989 feature_gate.go:330] unrecognized feature gate: Example Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684430 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684437 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684442 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684448 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684454 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684459 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684464 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684470 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684477 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684487 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684494 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684500 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684506 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684512 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684519 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684527 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684535 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684542 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684549 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684557 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684563 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684570 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684577 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684583 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684589 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684598 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684605 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684612 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684619 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684625 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684633 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684641 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684648 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684684 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684691 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684703 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684709 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684716 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684722 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684727 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684732 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684739 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684744 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684749 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684754 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684759 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684764 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684768 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684773 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684778 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684784 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684790 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684797 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684803 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684812 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684820 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684828 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684835 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684842 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684848 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684855 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684862 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684869 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684874 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684879 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684884 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684890 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684897 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684902 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.684907 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.684915 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.693424 4989 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.693467 4989 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693575 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693586 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693594 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693602 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693608 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693614 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693621 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693628 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693634 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693641 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693646 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693671 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693677 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693683 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693690 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693696 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693702 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693708 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693714 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693724 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693733 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693740 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693747 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693753 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693760 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693767 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693773 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693779 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693785 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693791 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693797 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693803 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693808 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693814 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693821 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693826 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693831 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693836 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693841 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693846 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693850 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693855 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693860 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693865 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693870 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693875 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693879 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693884 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693889 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693893 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693900 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693906 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693912 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693917 4989 feature_gate.go:330] unrecognized feature gate: Example Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693922 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693927 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693932 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693937 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693941 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693946 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693951 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693958 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693964 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693969 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693974 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693980 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693986 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693992 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.693997 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694002 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694007 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.694017 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694180 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694187 4989 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694193 4989 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694198 4989 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694204 4989 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694209 4989 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694214 4989 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694219 4989 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694224 4989 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694229 4989 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694234 4989 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694239 4989 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694244 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694249 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694256 4989 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694263 4989 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694269 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694276 4989 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694282 4989 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694287 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694292 4989 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694297 4989 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694304 4989 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694309 4989 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694315 4989 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694320 4989 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694325 4989 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694330 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694334 4989 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694340 4989 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694346 4989 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694352 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694357 4989 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694364 4989 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694371 4989 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694377 4989 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694382 4989 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694387 4989 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694392 4989 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694397 4989 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694402 4989 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694407 4989 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694412 4989 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694417 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694422 4989 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694427 4989 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694432 4989 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694436 4989 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694442 4989 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694447 4989 feature_gate.go:330] unrecognized feature gate: Example Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694452 4989 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694457 4989 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694462 4989 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694466 4989 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694472 4989 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694477 4989 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694482 4989 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694487 4989 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694492 4989 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694497 4989 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694502 4989 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694507 4989 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694512 4989 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694517 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694522 4989 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694527 4989 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694532 4989 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694537 4989 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694542 4989 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694548 4989 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.694554 4989 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.694562 4989 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.694789 4989 server.go:940] "Client rotation is on, will bootstrap in background" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.699332 4989 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.699418 4989 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.701180 4989 server.go:997] "Starting client certificate rotation" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.701206 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.701443 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 07:20:23.708242338 +0000 UTC Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.701533 4989 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1534h41m14.00671447s for next certificate rotation Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.725904 4989 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.727818 4989 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.742519 4989 log.go:25] "Validated CRI v1 runtime API" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.782955 4989 log.go:25] "Validated CRI v1 image API" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.787961 4989 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.795103 4989 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-06-08-33-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.795149 4989 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.831569 4989 manager.go:217] Machine: {Timestamp:2025-10-06 08:39:09.827295072 +0000 UTC m=+0.617320722 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:ef74357a-4028-4330-ae84-d5f17f42c6f5 BootID:98580f71-44d1-4b79-83d9-4c80b9cd1cfc Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3e:ed:ac Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3e:ed:ac Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:63:ce:24 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:74:b8:5a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:88:12:31 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2a:d4:8a Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:bb:bf:59 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:77:ae:94:c8:5a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:42:d4:22:6b:d1:df Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.832030 4989 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.832318 4989 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.833916 4989 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.834231 4989 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.834285 4989 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.834586 4989 topology_manager.go:138] "Creating topology manager with none policy" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.834605 4989 container_manager_linux.go:303] "Creating device plugin manager" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.835398 4989 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.835449 4989 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.836240 4989 state_mem.go:36] "Initialized new in-memory state store" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.836421 4989 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.844864 4989 kubelet.go:418] "Attempting to sync node with API server" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.844913 4989 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.844947 4989 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.844968 4989 kubelet.go:324] "Adding apiserver pod source" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.844986 4989 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.850313 4989 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.851449 4989 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.854504 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.854620 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.854646 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.854808 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.855365 4989 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857067 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857112 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857129 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857143 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857166 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857184 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857198 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857222 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857237 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857275 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857307 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.857321 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.858726 4989 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.859288 4989 server.go:1280] "Started kubelet" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.860458 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.861278 4989 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.861285 4989 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 06 08:39:09 crc systemd[1]: Started Kubernetes Kubelet. Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.861983 4989 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.864350 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.864398 4989 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.864588 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 04:26:50.483442669 +0000 UTC Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.864641 4989 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1387h47m40.618805134s for next certificate rotation Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.864749 4989 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.869523 4989 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.871718 4989 server.go:460] "Adding debug handlers to kubelet server" Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.873093 4989 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.873468 4989 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.873874 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.23:6443: connect: connection refused" interval="200ms" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.874415 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.874507 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.878093 4989 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.878537 4989 factory.go:55] Registering systemd factory Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.878556 4989 factory.go:221] Registration of the systemd container factory successfully Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.879147 4989 factory.go:153] Registering CRI-O factory Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.879178 4989 factory.go:221] Registration of the crio container factory successfully Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.879215 4989 factory.go:103] Registering Raw factory Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.879255 4989 manager.go:1196] Started watching for new ooms in manager Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.885673 4989 manager.go:319] Starting recovery of all containers Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.884924 4989 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.23:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186bda29306cbee2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-06 08:39:09.859258082 +0000 UTC m=+0.649283672,LastTimestamp:2025-10-06 08:39:09.859258082 +0000 UTC m=+0.649283672,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887754 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887808 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887823 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887835 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887847 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887858 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887870 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887882 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887895 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887904 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887943 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887952 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887964 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887975 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887983 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.887991 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888000 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888008 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888017 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888026 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888035 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888044 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888053 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888062 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888106 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888117 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888129 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888140 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888149 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888158 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888168 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888176 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888185 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888193 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888202 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888213 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888225 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888236 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888247 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888258 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888270 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888281 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888293 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888306 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888318 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888329 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888340 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888385 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888397 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888408 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888420 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888431 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888444 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888454 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888463 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888472 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888481 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888490 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888498 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888508 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.888516 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890287 4989 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890308 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890320 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890330 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890340 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890350 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890359 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890367 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890379 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890388 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890397 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890408 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890418 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890427 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890436 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890445 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890454 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890462 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890470 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890480 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890489 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890498 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890532 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890541 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890550 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890560 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890569 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890578 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890588 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890597 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890607 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890615 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890624 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890632 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890640 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890647 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890716 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890725 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890733 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890742 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890751 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890759 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890769 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890778 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890791 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890800 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890811 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890821 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890830 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890840 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890849 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890858 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890867 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890877 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890886 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890896 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890906 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890916 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890925 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890934 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890942 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890952 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890964 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890973 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890989 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.890999 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891009 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891018 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891028 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891037 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891055 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891064 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891074 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891084 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891093 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891102 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891111 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891121 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891131 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891139 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891148 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891157 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891165 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891174 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891182 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891191 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891198 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891207 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891216 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891224 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891233 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891241 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891249 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891452 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891464 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891472 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891482 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891490 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891499 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891508 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891516 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891525 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891534 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891542 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891550 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891559 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891568 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891576 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891584 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891594 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891603 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891612 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891621 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891631 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891640 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891661 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891670 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891678 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891688 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891696 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891706 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891716 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891724 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891733 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891741 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891750 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891761 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891769 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891778 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891786 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891794 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891803 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891812 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891820 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891829 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891838 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891847 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891856 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891864 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891872 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891880 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891890 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891898 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891907 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891916 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891924 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891933 4989 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891941 4989 reconstruct.go:97] "Volume reconstruction finished" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.891947 4989 reconciler.go:26] "Reconciler: start to sync state" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.905912 4989 manager.go:324] Recovery completed Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.919888 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.922378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.922432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.922448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.924101 4989 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.924131 4989 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.924152 4989 state_mem.go:36] "Initialized new in-memory state store" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.932895 4989 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.934443 4989 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.934480 4989 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.934691 4989 kubelet.go:2335] "Starting kubelet main sync loop" Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.934752 4989 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 06 08:39:09 crc kubenswrapper[4989]: W1006 08:39:09.935328 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.935391 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.936943 4989 policy_none.go:49] "None policy: Start" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.937737 4989 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.937765 4989 state_mem.go:35] "Initializing new in-memory state store" Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.973886 4989 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.982582 4989 manager.go:334] "Starting Device Plugin manager" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.982697 4989 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.982717 4989 server.go:79] "Starting device plugin registration server" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.983118 4989 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.983137 4989 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.983319 4989 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.983407 4989 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 06 08:39:09 crc kubenswrapper[4989]: I1006 08:39:09.983422 4989 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 06 08:39:09 crc kubenswrapper[4989]: E1006 08:39:09.990188 4989 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.034895 4989 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.034972 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.035713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.035739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.035747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.035838 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.036200 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.036231 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037187 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037366 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037490 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.037517 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038382 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038398 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.038680 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039223 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039331 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039355 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039940 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.039959 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.040770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.074311 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.23:6443: connect: connection refused" interval="400ms" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.084087 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.085339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.085379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.085390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.085413 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.085965 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.23:6443: connect: connection refused" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093508 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093547 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093582 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093613 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093644 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093705 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093739 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093770 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093802 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093830 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093885 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093950 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.093992 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.094053 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195236 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195307 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195373 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195405 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195435 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195462 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195495 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195500 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195534 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195544 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195608 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195687 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195707 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195573 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195744 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.196190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.196244 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195733 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.196309 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195710 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.195736 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.196423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.197871 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.197935 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.197984 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.198504 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.198569 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.198613 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.287012 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.289149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.289215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.289238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.289284 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.290525 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.23:6443: connect: connection refused" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.366115 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.372293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.386818 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.402509 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.407352 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.413561 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-bd7ee7e5a96640d6e985439ffe240d879e06ebc6f4a3af838593b77d6202630c WatchSource:0}: Error finding container bd7ee7e5a96640d6e985439ffe240d879e06ebc6f4a3af838593b77d6202630c: Status 404 returned error can't find the container with id bd7ee7e5a96640d6e985439ffe240d879e06ebc6f4a3af838593b77d6202630c Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.418756 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6a0be3fa8d28209ae3f1a7efd843cc7804942d42ada23256525583ce0c823aac WatchSource:0}: Error finding container 6a0be3fa8d28209ae3f1a7efd843cc7804942d42ada23256525583ce0c823aac: Status 404 returned error can't find the container with id 6a0be3fa8d28209ae3f1a7efd843cc7804942d42ada23256525583ce0c823aac Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.428763 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-1c712a57f3326ea05266097ac9716dd24a672f661b37969683573a9c32376cd8 WatchSource:0}: Error finding container 1c712a57f3326ea05266097ac9716dd24a672f661b37969683573a9c32376cd8: Status 404 returned error can't find the container with id 1c712a57f3326ea05266097ac9716dd24a672f661b37969683573a9c32376cd8 Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.440831 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-781b4a39bcf12f3979a9af8252064cfed820ba53c10b3a474c1e22292b09f114 WatchSource:0}: Error finding container 781b4a39bcf12f3979a9af8252064cfed820ba53c10b3a474c1e22292b09f114: Status 404 returned error can't find the container with id 781b4a39bcf12f3979a9af8252064cfed820ba53c10b3a474c1e22292b09f114 Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.442112 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2662a8b53a22bf2a8422fc9f70421f3b8f66a6746c4582ebc6340b185d61710d WatchSource:0}: Error finding container 2662a8b53a22bf2a8422fc9f70421f3b8f66a6746c4582ebc6340b185d61710d: Status 404 returned error can't find the container with id 2662a8b53a22bf2a8422fc9f70421f3b8f66a6746c4582ebc6340b185d61710d Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.475203 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.23:6443: connect: connection refused" interval="800ms" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.691520 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.693112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.693151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.693162 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.693186 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.693620 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.23:6443: connect: connection refused" node="crc" Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.753409 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.753510 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.861497 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:10 crc kubenswrapper[4989]: W1006 08:39:10.890679 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:10 crc kubenswrapper[4989]: E1006 08:39:10.890847 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.942104 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a0be3fa8d28209ae3f1a7efd843cc7804942d42ada23256525583ce0c823aac"} Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.944034 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bd7ee7e5a96640d6e985439ffe240d879e06ebc6f4a3af838593b77d6202630c"} Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.945444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2662a8b53a22bf2a8422fc9f70421f3b8f66a6746c4582ebc6340b185d61710d"} Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.946172 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c712a57f3326ea05266097ac9716dd24a672f661b37969683573a9c32376cd8"} Oct 06 08:39:10 crc kubenswrapper[4989]: I1006 08:39:10.946786 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"781b4a39bcf12f3979a9af8252064cfed820ba53c10b3a474c1e22292b09f114"} Oct 06 08:39:11 crc kubenswrapper[4989]: W1006 08:39:11.022557 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:11 crc kubenswrapper[4989]: E1006 08:39:11.022621 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:11 crc kubenswrapper[4989]: W1006 08:39:11.266113 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:11 crc kubenswrapper[4989]: E1006 08:39:11.266223 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:11 crc kubenswrapper[4989]: E1006 08:39:11.276493 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.23:6443: connect: connection refused" interval="1.6s" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.493933 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.499343 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.499401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.499420 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.499458 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:11 crc kubenswrapper[4989]: E1006 08:39:11.500295 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.23:6443: connect: connection refused" node="crc" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.861890 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.955101 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.955059 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.955283 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.955340 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.955373 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.956262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.956293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.956304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.957777 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7" exitCode=0 Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.957846 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.957918 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.960784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.960858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.960885 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.961977 4989 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0" exitCode=0 Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.962021 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.962188 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.963971 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.964408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.964424 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.964432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965033 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965293 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568" exitCode=0 Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965390 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.965521 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.966681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.966703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.966712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.967857 4989 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c" exitCode=0 Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.967894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c"} Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.967948 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.968996 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.969015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:11 crc kubenswrapper[4989]: I1006 08:39:11.969023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:12 crc kubenswrapper[4989]: W1006 08:39:12.547338 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:12 crc kubenswrapper[4989]: E1006 08:39:12.547448 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.861235 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:12 crc kubenswrapper[4989]: E1006 08:39:12.877681 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.23:6443: connect: connection refused" interval="3.2s" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.971977 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.971986 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.973409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.973457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.973475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.976914 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.976942 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.976953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.976961 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.980633 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.980680 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.980694 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.980775 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.981845 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.981875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.981886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.983608 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576" exitCode=0 Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.983751 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984089 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984390 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576"} Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:12 crc kubenswrapper[4989]: I1006 08:39:12.984959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:13 crc kubenswrapper[4989]: W1006 08:39:13.032808 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:13 crc kubenswrapper[4989]: E1006 08:39:13.032929 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.100600 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.101912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.101940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.101949 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.101974 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:13 crc kubenswrapper[4989]: E1006 08:39:13.102420 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.23:6443: connect: connection refused" node="crc" Oct 06 08:39:13 crc kubenswrapper[4989]: W1006 08:39:13.254414 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.23:6443: connect: connection refused Oct 06 08:39:13 crc kubenswrapper[4989]: E1006 08:39:13.254508 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.23:6443: connect: connection refused" logger="UnhandledError" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.526527 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.990985 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0"} Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.991074 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.992810 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.992893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.992913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.994853 4989 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b" exitCode=0 Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.994939 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.994947 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b"} Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.994971 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.995008 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.995064 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.995728 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.996946 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.997000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.997021 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.997995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.998051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.998073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.998917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.998957 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:13 crc kubenswrapper[4989]: I1006 08:39:13.998973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:14 crc kubenswrapper[4989]: I1006 08:39:14.001097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:14 crc kubenswrapper[4989]: I1006 08:39:14.001144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:14 crc kubenswrapper[4989]: I1006 08:39:14.001160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:14 crc kubenswrapper[4989]: I1006 08:39:14.260852 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:14 crc kubenswrapper[4989]: I1006 08:39:14.867794 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.001978 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191"} Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33"} Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002059 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab"} Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002077 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21"} Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002009 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002132 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.002091 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:15 crc kubenswrapper[4989]: I1006 08:39:15.003699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.010392 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.010459 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.011200 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.011535 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b"} Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012095 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.012312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.303577 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.305610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.305707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.305741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.305783 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.707310 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.946325 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.946600 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.948149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.948189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:16 crc kubenswrapper[4989]: I1006 08:39:16.948227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.013231 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.014438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.014481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.014492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.208758 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.209001 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.210345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.210385 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.210396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.527470 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.527739 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.527794 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.529289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.529358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.529374 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.868047 4989 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 08:39:17 crc kubenswrapper[4989]: I1006 08:39:17.868167 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 08:39:18 crc kubenswrapper[4989]: I1006 08:39:18.015369 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:18 crc kubenswrapper[4989]: I1006 08:39:18.016340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:18 crc kubenswrapper[4989]: I1006 08:39:18.016376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:18 crc kubenswrapper[4989]: I1006 08:39:18.016386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:19 crc kubenswrapper[4989]: I1006 08:39:19.918384 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:19 crc kubenswrapper[4989]: I1006 08:39:19.918623 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:19 crc kubenswrapper[4989]: I1006 08:39:19.920206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:19 crc kubenswrapper[4989]: I1006 08:39:19.920256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:19 crc kubenswrapper[4989]: I1006 08:39:19.920273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:19 crc kubenswrapper[4989]: E1006 08:39:19.990373 4989 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.635334 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.635478 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.636721 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.636752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.636761 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:21 crc kubenswrapper[4989]: I1006 08:39:21.644866 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:22 crc kubenswrapper[4989]: I1006 08:39:22.024356 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:22 crc kubenswrapper[4989]: I1006 08:39:22.025537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:22 crc kubenswrapper[4989]: I1006 08:39:22.025589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:22 crc kubenswrapper[4989]: I1006 08:39:22.025606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:22 crc kubenswrapper[4989]: I1006 08:39:22.031749 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.027145 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.028601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.028647 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.028690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.546949 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.547171 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.548390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.548431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.548441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.862405 4989 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 06 08:39:23 crc kubenswrapper[4989]: W1006 08:39:23.935150 4989 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 06 08:39:23 crc kubenswrapper[4989]: I1006 08:39:23.935251 4989 trace.go:236] Trace[2129724083]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 08:39:13.934) (total time: 10000ms): Oct 06 08:39:23 crc kubenswrapper[4989]: Trace[2129724083]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (08:39:23.935) Oct 06 08:39:23 crc kubenswrapper[4989]: Trace[2129724083]: [10.000791144s] [10.000791144s] END Oct 06 08:39:23 crc kubenswrapper[4989]: E1006 08:39:23.935276 4989 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.030496 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.032038 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0" exitCode=255 Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.032094 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0"} Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.032303 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.033387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.033440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.033457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.034215 4989 scope.go:117] "RemoveContainer" containerID="b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.261197 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.261374 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.365781 4989 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 06 08:39:24 crc kubenswrapper[4989]: I1006 08:39:24.365862 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.036096 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.039084 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440"} Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.039204 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.039942 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.039987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:25 crc kubenswrapper[4989]: I1006 08:39:25.039999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:27 crc kubenswrapper[4989]: I1006 08:39:27.868562 4989 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 06 08:39:27 crc kubenswrapper[4989]: I1006 08:39:27.868693 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 08:39:28 crc kubenswrapper[4989]: I1006 08:39:28.938859 4989 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.271114 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.271293 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.271595 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.275035 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.275110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.275135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.279094 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.365111 4989 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.367801 4989 trace.go:236] Trace[1445285454]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 08:39:17.005) (total time: 12362ms): Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[1445285454]: ---"Objects listed" error: 12362ms (08:39:29.367) Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[1445285454]: [12.362412094s] [12.362412094s] END Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.367856 4989 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.370192 4989 trace.go:236] Trace[301956426]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 08:39:17.732) (total time: 11637ms): Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[301956426]: ---"Objects listed" error: 11637ms (08:39:29.370) Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[301956426]: [11.637775634s] [11.637775634s] END Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.370227 4989 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.370233 4989 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.370636 4989 trace.go:236] Trace[221155928]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Oct-2025 08:39:17.599) (total time: 11771ms): Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[221155928]: ---"Objects listed" error: 11770ms (08:39:29.370) Oct 06 08:39:29 crc kubenswrapper[4989]: Trace[221155928]: [11.771034403s] [11.771034403s] END Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.370696 4989 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.371906 4989 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.858446 4989 apiserver.go:52] "Watching apiserver" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.863019 4989 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.863261 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.863935 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.863974 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.864150 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.864068 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.864242 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.864480 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.864777 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.865237 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.865300 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.871933 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.871988 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.872045 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.872154 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.872738 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.872799 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.872930 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.873256 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.873436 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.876332 4989 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.906048 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.918769 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.929143 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.939150 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.947944 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.957140 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.966934 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973459 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973491 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973512 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973529 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973545 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973559 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973575 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973592 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973605 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973620 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973634 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973680 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973702 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973718 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973732 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973759 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973781 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973796 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973810 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973824 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973838 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973854 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973871 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973893 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973907 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973923 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973937 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973958 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973973 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.973987 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974001 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974020 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974035 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974053 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974068 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974082 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974096 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974111 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974135 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974151 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974165 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974193 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974208 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974223 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974239 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974254 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974270 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974284 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974299 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974314 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974328 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974344 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974358 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974375 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974392 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974408 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974422 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974438 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974452 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974490 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974507 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974522 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974542 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974556 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974571 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974588 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974609 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974632 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974647 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974690 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974706 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974726 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974740 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974771 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974785 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974800 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974815 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974832 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974864 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974879 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974894 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974909 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974925 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974939 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974953 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974968 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.974984 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975001 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975016 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975031 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975045 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975061 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975078 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975092 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975108 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975123 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975138 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975152 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975169 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975184 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975199 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975214 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975230 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975248 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975263 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975279 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975309 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975323 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975340 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975364 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975380 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975396 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975411 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975426 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975441 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975456 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975487 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975508 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975579 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975597 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975615 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975633 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975662 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975685 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975702 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975718 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975733 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975749 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975781 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975796 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975817 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975832 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975865 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975882 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975898 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975913 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975930 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975949 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975965 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975981 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.975997 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976013 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976028 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976043 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976058 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976075 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976091 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976107 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976122 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976138 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976172 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976188 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976204 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976222 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976237 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976253 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976269 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976302 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976318 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976333 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976351 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976372 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976391 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976406 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976423 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976441 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976458 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976475 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976509 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976592 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976611 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976626 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976681 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976699 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976715 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976734 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976750 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976782 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976799 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976818 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976836 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976851 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976871 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976918 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976943 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976970 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.976997 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977014 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977031 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977051 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977534 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977537 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977634 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977944 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977972 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978014 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.977072 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978041 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978078 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978145 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978171 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978216 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978104 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978316 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978336 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978355 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978374 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978391 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978540 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.978807 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979049 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979142 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979151 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979286 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979361 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979394 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.979554 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.980538 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981569 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981710 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981836 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981905 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981923 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981930 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982035 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.981986 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982272 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982370 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982394 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982404 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982700 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982607 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982597 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982748 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.982816 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.983354 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.983713 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.983884 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.984010 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985099 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.984981 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985392 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985470 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985505 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985587 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985826 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985886 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985962 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.985977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.986140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.986235 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.986293 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.986310 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.986347 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.987961 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.987976 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.988255 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.988418 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.988425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.988456 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.988729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.989121 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.989219 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:30.489173458 +0000 UTC m=+21.279199038 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.997478 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.997819 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.997995 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.998203 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: E1006 08:39:29.998310 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.998374 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.998502 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.998956 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999153 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999169 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999402 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999563 4989 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999576 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999784 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:29 crc kubenswrapper[4989]: I1006 08:39:29.999870 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:29.999993 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000072 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000258 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000273 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000484 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000521 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000613 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000612 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.000977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.001598 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.001684 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.001731 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.001826 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002036 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:29.998817 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002089 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002181 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002232 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002323 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002719 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.002900 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003070 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003156 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003229 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003164 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003555 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003756 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003831 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003908 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003954 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.003956 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.004103 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.004456 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.004538 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.004956 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.004976 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.005069 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.005771 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.005770 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006063 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006049 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006402 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006751 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.006947 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007049 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007357 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007426 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007093 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007731 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.007828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008291 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008486 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008545 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008691 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.008941 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.009007 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.010135 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.010833 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.011112 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.011294 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.011305 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.011937 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012351 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012397 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012530 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012845 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.012994 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013113 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013136 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013409 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013692 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013885 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.013975 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.014162 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.014553 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.014637 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.014676 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015099 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015128 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015170 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015389 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015520 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015574 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.015946 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.016018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.016089 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.018361 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018388 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.018403 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.016462 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.016907 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.017105 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018502 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.017236 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.017261 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.017760 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018133 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.017921 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.016137 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:30.516111856 +0000 UTC m=+21.306137436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018765 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.018904 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.019187 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.019404 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.019993 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.020186 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.021770 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.024124 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.024125 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.032194 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.032280 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.032569 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.032639 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.032939 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.033277 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.033571 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.033593 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.033608 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.033865 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.033961 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:30.533944548 +0000 UTC m=+21.323970218 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.034032 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:30.53402366 +0000 UTC m=+21.324049240 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.034246 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:39:30.534233435 +0000 UTC m=+21.324259085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034282 4989 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034843 4989 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034916 4989 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034974 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035040 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035097 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035156 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035211 4989 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035264 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035326 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035385 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.035439 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.034931 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.041172 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.041756 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.046030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.047960 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.069393 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.071103 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.074184 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.076835 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.086194 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.091535 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.099400 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.109307 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.117866 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136020 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136060 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136152 4989 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136163 4989 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136174 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136182 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136190 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136199 4989 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136208 4989 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136218 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136229 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136237 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136246 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136255 4989 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136257 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136266 4989 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136319 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136332 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136346 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136360 4989 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136375 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136387 4989 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136401 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136196 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136887 4989 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136903 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136914 4989 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136926 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136937 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136949 4989 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136960 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136971 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136981 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.136993 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137006 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137017 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137028 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137038 4989 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137047 4989 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137055 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137064 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137071 4989 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137080 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137088 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137096 4989 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137105 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137116 4989 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137127 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137138 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137151 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137161 4989 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137172 4989 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137183 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137193 4989 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137204 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137215 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137228 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137239 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137251 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137263 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137274 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137285 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137297 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137307 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137318 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137328 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137340 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137350 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137363 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137377 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137391 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137403 4989 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137414 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137425 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137437 4989 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137449 4989 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137461 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137471 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137485 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137497 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137509 4989 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137521 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137532 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137544 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137555 4989 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137566 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137578 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137590 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137601 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137612 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137623 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137635 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137646 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137676 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137687 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137699 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137710 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137723 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137734 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137745 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137758 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137770 4989 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137782 4989 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137794 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137806 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137817 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137828 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137859 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137870 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137881 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137894 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137907 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137919 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137929 4989 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137939 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137949 4989 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137959 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.137970 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138387 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138433 4989 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138445 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138455 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138465 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138515 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138527 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138538 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138548 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138560 4989 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138572 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138585 4989 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138599 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138609 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138620 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.138630 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140357 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140386 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140399 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140460 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140473 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140485 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140673 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.140698 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141325 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141391 4989 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141406 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141449 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141462 4989 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141472 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141484 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141495 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141506 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141516 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141526 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141536 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141546 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141556 4989 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141566 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141577 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141589 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141598 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141610 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141621 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141632 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141641 4989 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141667 4989 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141678 4989 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141690 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141702 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141713 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141723 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141734 4989 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141745 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141758 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141770 4989 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141782 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141794 4989 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141805 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141816 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141827 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141838 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141849 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141860 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141871 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141882 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141892 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.141903 4989 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.181867 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.193540 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.204053 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 06 08:39:30 crc kubenswrapper[4989]: W1006 08:39:30.205479 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-835b333dd583397c252aaba2a816cad30186e60913b383e27e667bb91d18f8d4 WatchSource:0}: Error finding container 835b333dd583397c252aaba2a816cad30186e60913b383e27e667bb91d18f8d4: Status 404 returned error can't find the container with id 835b333dd583397c252aaba2a816cad30186e60913b383e27e667bb91d18f8d4 Oct 06 08:39:30 crc kubenswrapper[4989]: W1006 08:39:30.214336 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a0ea0c764deab1b3b9898ffba4b2f08dedca324d91e950573356c46262cd946d WatchSource:0}: Error finding container a0ea0c764deab1b3b9898ffba4b2f08dedca324d91e950573356c46262cd946d: Status 404 returned error can't find the container with id a0ea0c764deab1b3b9898ffba4b2f08dedca324d91e950573356c46262cd946d Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.545228 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.545449 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:39:31.545419477 +0000 UTC m=+22.335445067 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.545635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.545794 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.545836 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.545861 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:31.545847767 +0000 UTC m=+22.335873397 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.545894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.545949 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.545987 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546004 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546017 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546033 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546046 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546060 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:31.546047202 +0000 UTC m=+22.336072782 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546063 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546080 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:31.546068683 +0000 UTC m=+22.336094273 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546082 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.546192 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:31.546181116 +0000 UTC m=+22.336206706 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:30 crc kubenswrapper[4989]: I1006 08:39:30.935835 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:30 crc kubenswrapper[4989]: E1006 08:39:30.935984 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.058937 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a0ea0c764deab1b3b9898ffba4b2f08dedca324d91e950573356c46262cd946d"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.060730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.060754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.060763 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"835b333dd583397c252aaba2a816cad30186e60913b383e27e667bb91d18f8d4"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.062145 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.062198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7ee2e201d53a2b55f1355bdd751113ea8bcca935affbead5fa9778b36c72b4e6"} Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.074160 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.084330 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.093962 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.103964 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.115717 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.128534 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.136376 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.150353 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.159725 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.168369 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.177189 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.183802 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.192489 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.215323 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.553329 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.553458 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553540 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:39:33.553498366 +0000 UTC m=+24.343523986 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.553599 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.553713 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553742 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.553756 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553878 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553888 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553918 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553937 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553963 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:33.553936057 +0000 UTC m=+24.343961677 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553876 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553990 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.553996 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:33.553980058 +0000 UTC m=+24.344005688 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.554002 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.554049 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:33.554034939 +0000 UTC m=+24.344060549 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.554118 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:33.55406068 +0000 UTC m=+24.344086310 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.935592 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.935728 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.935812 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:31 crc kubenswrapper[4989]: E1006 08:39:31.936045 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.939782 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.940449 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.941251 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.941851 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.942386 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.942896 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.943437 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.943971 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.944544 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.946531 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.947016 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.948002 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.948505 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.949004 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.949842 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.950346 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.951206 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.951554 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.952102 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.953035 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.953442 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.954351 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.954789 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.956033 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.957782 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.958717 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.959925 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.960493 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.961670 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.962241 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.963399 4989 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.963515 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.965087 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.965536 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.966329 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.967752 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.968331 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.969331 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.969942 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.970910 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.971355 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.972266 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.972908 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.973890 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.974380 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.975228 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.975754 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.976758 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.977210 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.978004 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.978451 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.979310 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.979881 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 06 08:39:31 crc kubenswrapper[4989]: I1006 08:39:31.980332 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 06 08:39:32 crc kubenswrapper[4989]: I1006 08:39:32.935548 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:32 crc kubenswrapper[4989]: E1006 08:39:32.935744 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.071157 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127"} Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.092625 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.109466 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.131498 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.146678 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.157151 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.169183 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.180286 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.570981 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.571039 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.571064 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.571083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.571099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571162 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571205 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:37.5711929 +0000 UTC m=+28.361218480 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571480 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:39:37.571470787 +0000 UTC m=+28.361496367 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571667 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571752 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571786 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:37.571777535 +0000 UTC m=+28.361803105 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571714 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571855 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571871 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.571914 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:37.571902108 +0000 UTC m=+28.361927688 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.572068 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.572143 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.572247 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:37.572236516 +0000 UTC m=+28.362262096 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.586263 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.610127 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.610771 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.635947 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.652106 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.660591 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7b8zg"] Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.660813 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-ddvqd"] Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.660962 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.661197 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.663245 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.663409 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.663526 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.663994 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.664148 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.664617 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.666864 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.667020 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.679890 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.690425 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.701236 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.711944 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.729124 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.740329 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.752614 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.756000 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.767894 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.771948 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6pv\" (UniqueName: \"kubernetes.io/projected/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-kube-api-access-wb6pv\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.772016 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-serviceca\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.772051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-host\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.772072 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-hosts-file\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.772100 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcb2j\" (UniqueName: \"kubernetes.io/projected/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-kube-api-access-xcb2j\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.782548 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.793921 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.804855 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.816315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:33Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-serviceca\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-host\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873230 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcb2j\" (UniqueName: \"kubernetes.io/projected/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-kube-api-access-xcb2j\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873256 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-hosts-file\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873289 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6pv\" (UniqueName: \"kubernetes.io/projected/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-kube-api-access-wb6pv\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-hosts-file\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.873498 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-host\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.876733 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-serviceca\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.900119 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6pv\" (UniqueName: \"kubernetes.io/projected/00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0-kube-api-access-wb6pv\") pod \"node-resolver-ddvqd\" (UID: \"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\") " pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.900451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcb2j\" (UniqueName: \"kubernetes.io/projected/70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77-kube-api-access-xcb2j\") pod \"node-ca-7b8zg\" (UID: \"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\") " pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.935150 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.935183 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.935301 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:33 crc kubenswrapper[4989]: E1006 08:39:33.935418 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.975328 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ddvqd" Oct 06 08:39:33 crc kubenswrapper[4989]: I1006 08:39:33.981151 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7b8zg" Oct 06 08:39:33 crc kubenswrapper[4989]: W1006 08:39:33.985272 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00c2dc1b_134d_4ee4_af3b_8143ed4ac2f0.slice/crio-791787154d2d7a171c98cfc351482e84c9ea23b03b16158315c5ae6e9e871618 WatchSource:0}: Error finding container 791787154d2d7a171c98cfc351482e84c9ea23b03b16158315c5ae6e9e871618: Status 404 returned error can't find the container with id 791787154d2d7a171c98cfc351482e84c9ea23b03b16158315c5ae6e9e871618 Oct 06 08:39:33 crc kubenswrapper[4989]: W1006 08:39:33.992109 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70b2c0ba_ddd3_41e6_ae65_5d0e9a4b3d77.slice/crio-17940f06d14d1e58769d3803e1a078b02cad17d6b6814c1af299d9d9564b79d2 WatchSource:0}: Error finding container 17940f06d14d1e58769d3803e1a078b02cad17d6b6814c1af299d9d9564b79d2: Status 404 returned error can't find the container with id 17940f06d14d1e58769d3803e1a078b02cad17d6b6814c1af299d9d9564b79d2 Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.080271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7b8zg" event={"ID":"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77","Type":"ContainerStarted","Data":"17940f06d14d1e58769d3803e1a078b02cad17d6b6814c1af299d9d9564b79d2"} Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.081999 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ddvqd" event={"ID":"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0","Type":"ContainerStarted","Data":"791787154d2d7a171c98cfc351482e84c9ea23b03b16158315c5ae6e9e871618"} Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.542006 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-wz8dx"] Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.542335 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.542611 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-v6vsq"] Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.542968 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5s7m9"] Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.543057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.543815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: W1006 08:39:34.544312 4989 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 06 08:39:34 crc kubenswrapper[4989]: E1006 08:39:34.544355 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 06 08:39:34 crc kubenswrapper[4989]: W1006 08:39:34.544319 4989 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 06 08:39:34 crc kubenswrapper[4989]: E1006 08:39:34.544394 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 06 08:39:34 crc kubenswrapper[4989]: W1006 08:39:34.544323 4989 reflector.go:561] object-"openshift-multus"/"multus-daemon-config": failed to list *v1.ConfigMap: configmaps "multus-daemon-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 06 08:39:34 crc kubenswrapper[4989]: E1006 08:39:34.544418 4989 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-daemon-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"multus-daemon-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.545453 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.545865 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.547407 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.547478 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.547702 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.549812 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t6mdb"] Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.552000 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.552005 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.552406 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.555611 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.555699 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.555755 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.555877 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.555961 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.559176 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.559296 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.559449 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.559635 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.598584 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.624444 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.666787 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680399 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680444 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680514 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nl6q\" (UniqueName: \"kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680570 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-socket-dir-parent\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680596 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-conf-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680620 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcqm5\" (UniqueName: \"kubernetes.io/projected/e941df51-dd85-47d8-bb33-ab46f2c13e3c-kube-api-access-hcqm5\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680645 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-proxy-tls\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680684 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-system-cni-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680703 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680732 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680754 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680803 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-mcd-auth-proxy-config\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680838 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680860 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680880 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680902 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680945 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-hostroot\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680967 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sh5x\" (UniqueName: \"kubernetes.io/projected/b901f1ca-c342-4faa-a715-0ff3d94bc369-kube-api-access-4sh5x\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.680999 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681046 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-cnibin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681072 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-multus-certs\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681092 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-etc-kubernetes\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681129 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681164 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681188 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-k8s-cni-cncf-io\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681208 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-netns\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681228 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz8hg\" (UniqueName: \"kubernetes.io/projected/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-kube-api-access-rz8hg\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-multus\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681283 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cnibin\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681302 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681322 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681344 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681366 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-os-release\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681383 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681402 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681442 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681463 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-system-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681484 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681505 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681525 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-rootfs\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681579 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-bin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681600 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-os-release\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681621 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681642 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-kubelet\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681680 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.681701 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.709011 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.748549 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.774717 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782590 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-mcd-auth-proxy-config\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782664 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782711 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782734 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782762 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sh5x\" (UniqueName: \"kubernetes.io/projected/b901f1ca-c342-4faa-a715-0ff3d94bc369-kube-api-access-4sh5x\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782786 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782815 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-hostroot\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782835 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782854 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-cnibin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782872 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-multus-certs\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782891 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-etc-kubernetes\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782914 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782937 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-k8s-cni-cncf-io\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-netns\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.782979 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz8hg\" (UniqueName: \"kubernetes.io/projected/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-kube-api-access-rz8hg\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-multus\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783035 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cnibin\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783057 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783076 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783120 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783141 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783163 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-os-release\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783218 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783239 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-rootfs\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783259 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-system-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783279 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783316 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-bin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783362 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-os-release\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783381 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783400 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-kubelet\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783419 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783441 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783460 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783498 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nl6q\" (UniqueName: \"kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783520 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-socket-dir-parent\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-conf-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783559 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcqm5\" (UniqueName: \"kubernetes.io/projected/e941df51-dd85-47d8-bb33-ab46f2c13e3c-kube-api-access-hcqm5\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783578 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-proxy-tls\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783598 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-system-cni-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783636 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783677 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.783754 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-mcd-auth-proxy-config\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784611 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784675 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784726 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784808 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-rootfs\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784880 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-system-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784929 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784950 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.784971 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-hostroot\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785628 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785714 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-cnibin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785751 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-multus-certs\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785785 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-etc-kubernetes\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785817 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785840 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785854 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-k8s-cni-cncf-io\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785887 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-run-netns\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.785888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-bin\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786084 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-cni-multus\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786123 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cnibin\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786139 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-os-release\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786158 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786176 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-cni-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-host-var-lib-kubelet\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-os-release\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786252 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786177 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786262 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-socket-dir-parent\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786286 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786251 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786324 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-conf-dir\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786601 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786627 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-system-cni-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786645 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786794 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e941df51-dd85-47d8-bb33-ab46f2c13e3c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.786798 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.787026 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.791108 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-proxy-tls\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.792947 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.793098 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.803973 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sh5x\" (UniqueName: \"kubernetes.io/projected/b901f1ca-c342-4faa-a715-0ff3d94bc369-kube-api-access-4sh5x\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.805786 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.805923 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz8hg\" (UniqueName: \"kubernetes.io/projected/f78ef9f2-a706-4a6c-8a9f-b232f97e9211-kube-api-access-rz8hg\") pod \"machine-config-daemon-v6vsq\" (UID: \"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\") " pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.807590 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcqm5\" (UniqueName: \"kubernetes.io/projected/e941df51-dd85-47d8-bb33-ab46f2c13e3c-kube-api-access-hcqm5\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.809741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nl6q\" (UniqueName: \"kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q\") pod \"ovnkube-node-t6mdb\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.815831 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.826079 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.834572 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.844622 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.855097 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.865748 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.870901 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.874434 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.875647 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.877881 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.878218 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.890742 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.893920 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:34 crc kubenswrapper[4989]: W1006 08:39:34.906516 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91d0cca1_dfc6_4f12_a25a_7325782935ba.slice/crio-759d4d68f08c2d705bd6ef7f317cb1783cf09419ac82808310903662d1285dcd WatchSource:0}: Error finding container 759d4d68f08c2d705bd6ef7f317cb1783cf09419ac82808310903662d1285dcd: Status 404 returned error can't find the container with id 759d4d68f08c2d705bd6ef7f317cb1783cf09419ac82808310903662d1285dcd Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.912401 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.928179 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.935514 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:34 crc kubenswrapper[4989]: E1006 08:39:34.935613 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.940908 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.960448 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.977619 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.988567 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:34 crc kubenswrapper[4989]: I1006 08:39:34.998563 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:34Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.011562 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.023295 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.036445 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.060396 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.074261 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.084104 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7b8zg" event={"ID":"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77","Type":"ContainerStarted","Data":"eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.085086 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ddvqd" event={"ID":"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0","Type":"ContainerStarted","Data":"04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.086138 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" exitCode=0 Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.086161 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.086193 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"759d4d68f08c2d705bd6ef7f317cb1783cf09419ac82808310903662d1285dcd"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.088216 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.088268 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"9fcfd4501eff2613bf4f79cbf98fa375b0a2890855a43b8e3b890c7aa3002391"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.090075 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.105334 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.118853 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.129899 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.143059 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.155233 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.168243 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.184057 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.202215 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.215846 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.229372 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.238992 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.250576 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.263485 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.275297 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.293469 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.308344 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.322138 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.347938 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.359311 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.373944 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.374714 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.382589 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.397512 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.411518 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.422583 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.431946 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.445475 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.772202 4989 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.773982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.774019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.774029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.774109 4989 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.780222 4989 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.780470 4989 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.781500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.781563 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.781585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.781608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.781629 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.784839 4989 configmap.go:193] Couldn't get configMap openshift-multus/multus-daemon-config: failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.784856 4989 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.784933 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config podName:b901f1ca-c342-4faa-a715-0ff3d94bc369 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:36.284909845 +0000 UTC m=+27.074935445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "multus-daemon-config" (UniqueName: "kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config") pod "multus-wz8dx" (UID: "b901f1ca-c342-4faa-a715-0ff3d94bc369") : failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.785057 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy podName:b901f1ca-c342-4faa-a715-0ff3d94bc369 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:36.285013788 +0000 UTC m=+27.075039368 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy") pod "multus-wz8dx" (UID: "b901f1ca-c342-4faa-a715-0ff3d94bc369") : failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.786886 4989 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.786971 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy podName:e941df51-dd85-47d8-bb33-ab46f2c13e3c nodeName:}" failed. No retries permitted until 2025-10-06 08:39:36.286951116 +0000 UTC m=+27.076976736 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy") pod "multus-additional-cni-plugins-5s7m9" (UID: "e941df51-dd85-47d8-bb33-ab46f2c13e3c") : failed to sync configmap cache: timed out waiting for the condition Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.803892 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.807144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.807169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.807176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.807189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.807198 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.817622 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.824566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.824604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.824614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.824628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.824638 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.839083 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.842600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.842631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.842644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.842693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.842708 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.854778 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.857854 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.857886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.857896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.857908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.857918 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.869971 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:35Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.870084 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.872272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.872309 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.872320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.872336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.872347 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.930232 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.935366 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.935468 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.935379 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:35 crc kubenswrapper[4989]: E1006 08:39:35.935818 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.976548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.976882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.976897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.976914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:35 crc kubenswrapper[4989]: I1006 08:39:35.976926 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:35Z","lastTransitionTime":"2025-10-06T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.079727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.079813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.079833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.079860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.079875 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.088871 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.092032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095902 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095926 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095936 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.095962 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.106564 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.119991 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.128583 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.138515 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.147595 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.163157 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.174803 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.181956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.181979 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.181986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.181999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.182008 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.186460 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.198967 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.209610 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.221622 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.231760 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.241996 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.255366 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.276622 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:36Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.286319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.286568 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.286647 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.286760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.286829 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.296828 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.296890 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.296918 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.297569 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-cni-binary-copy\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.297870 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b901f1ca-c342-4faa-a715-0ff3d94bc369-multus-daemon-config\") pod \"multus-wz8dx\" (UID: \"b901f1ca-c342-4faa-a715-0ff3d94bc369\") " pod="openshift-multus/multus-wz8dx" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.298026 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e941df51-dd85-47d8-bb33-ab46f2c13e3c-cni-binary-copy\") pod \"multus-additional-cni-plugins-5s7m9\" (UID: \"e941df51-dd85-47d8-bb33-ab46f2c13e3c\") " pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.366328 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wz8dx" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.386687 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.388396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.388449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.388462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.388487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.388500 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: W1006 08:39:36.390673 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb901f1ca_c342_4faa_a715_0ff3d94bc369.slice/crio-e26f417ee7f74600d3582f0447618c8833fbb17dcb9b3fcb36847b8020255a2e WatchSource:0}: Error finding container e26f417ee7f74600d3582f0447618c8833fbb17dcb9b3fcb36847b8020255a2e: Status 404 returned error can't find the container with id e26f417ee7f74600d3582f0447618c8833fbb17dcb9b3fcb36847b8020255a2e Oct 06 08:39:36 crc kubenswrapper[4989]: W1006 08:39:36.398986 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode941df51_dd85_47d8_bb33_ab46f2c13e3c.slice/crio-88607f6ed0e1e19dd7fbfd02f18ef53ee9322499134fc609a7940d3125da8c38 WatchSource:0}: Error finding container 88607f6ed0e1e19dd7fbfd02f18ef53ee9322499134fc609a7940d3125da8c38: Status 404 returned error can't find the container with id 88607f6ed0e1e19dd7fbfd02f18ef53ee9322499134fc609a7940d3125da8c38 Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.491836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.492276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.492287 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.492301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.492311 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.594573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.594610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.594621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.594636 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.594645 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.696873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.696899 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.696908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.696922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.696931 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.800187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.800225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.800233 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.800247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.800258 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.903075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.903101 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.903109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.903122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.903129 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:36Z","lastTransitionTime":"2025-10-06T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:36 crc kubenswrapper[4989]: I1006 08:39:36.935037 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:36 crc kubenswrapper[4989]: E1006 08:39:36.935152 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.005208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.005244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.005255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.005268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.005279 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.099995 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerStarted","Data":"9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.100061 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerStarted","Data":"e26f417ee7f74600d3582f0447618c8833fbb17dcb9b3fcb36847b8020255a2e"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.106170 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb" exitCode=0 Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.106689 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.106723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerStarted","Data":"88607f6ed0e1e19dd7fbfd02f18ef53ee9322499134fc609a7940d3125da8c38"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.108113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.108161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.108180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.108202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.108217 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.124358 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.142692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.154871 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.169952 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.186721 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.197712 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.211136 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.212365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.212404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.212415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.212433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.212445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.221796 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.231631 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.251020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.267244 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.282664 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.293535 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.303149 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.314684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.314715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.314726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.314741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.314751 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.315077 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.325298 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.338645 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.351731 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.361710 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.379404 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.391540 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.402919 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.417006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.417038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.417046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.417060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.417069 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.418941 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.431265 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.440954 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.449377 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.459992 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.469584 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.478856 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.487287 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.519371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.519417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.519430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.519446 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.519462 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.610058 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.610184 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610234 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:39:45.610203251 +0000 UTC m=+36.400228831 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.610292 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610339 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.610353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.610397 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610362 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610457 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610466 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610502 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610521 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:45.610491448 +0000 UTC m=+36.400517048 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610529 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610594 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:45.61056922 +0000 UTC m=+36.400594920 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610476 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610648 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:45.610638051 +0000 UTC m=+36.400663741 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610431 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.610706 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:39:45.610697893 +0000 UTC m=+36.400723473 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.621811 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.622090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.622100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.622116 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.622124 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.725306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.725351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.725363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.725384 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.725413 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.828202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.828231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.828240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.828252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.828261 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.930024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.930054 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.930062 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.930076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.930085 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:37Z","lastTransitionTime":"2025-10-06T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.935482 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:37 crc kubenswrapper[4989]: I1006 08:39:37.935541 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.935623 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:37 crc kubenswrapper[4989]: E1006 08:39:37.936007 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.031891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.031926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.031937 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.031955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.031969 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.120306 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.123486 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5" exitCode=0 Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.123540 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.135607 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.135747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.135767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.135790 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.135807 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.144923 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.158486 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.170321 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.191227 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.205767 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.226849 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.238803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.238840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.238851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.238866 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.238876 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.245160 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.260106 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.275466 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.308739 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.328962 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.343474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.343502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.343510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.343525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.343535 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.351673 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.361095 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.372455 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.390337 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.446049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.446086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.446094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.446110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.446119 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.549150 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.549218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.549240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.549693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.549753 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.652071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.652125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.652141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.652164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.652183 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.754888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.754928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.754937 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.754952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.754962 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.859080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.859133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.859151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.859171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.859185 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.935601 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:38 crc kubenswrapper[4989]: E1006 08:39:38.935806 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.962205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.962251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.962262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.962276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:38 crc kubenswrapper[4989]: I1006 08:39:38.962286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:38Z","lastTransitionTime":"2025-10-06T08:39:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.064766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.064817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.064831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.064854 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.064868 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.130568 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09" exitCode=0 Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.130625 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.155564 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.171182 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.171216 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.171226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.171241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.171252 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.191688 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.209435 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.223968 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.239412 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.257262 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.270129 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.274361 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.274411 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.274420 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.274436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.274445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.281479 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.292933 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.303082 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.315282 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.336134 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.347474 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.356253 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.367983 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.381120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.381165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.381178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.381196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.381206 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.483636 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.483735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.483754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.483782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.483806 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.587495 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.587560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.587583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.587615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.587635 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.690342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.690408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.690431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.690460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.690482 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.792623 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.792679 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.792692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.792709 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.792726 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.895574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.895620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.895826 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.895851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.895866 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.922810 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.936095 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.936194 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:39 crc kubenswrapper[4989]: E1006 08:39:39.936323 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:39 crc kubenswrapper[4989]: E1006 08:39:39.936459 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.938407 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.952806 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.964554 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.980567 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.998032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.998066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.998080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.998097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:39 crc kubenswrapper[4989]: I1006 08:39:39.998108 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:39Z","lastTransitionTime":"2025-10-06T08:39:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.008448 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.025456 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.039818 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.052374 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.064692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.083574 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.096359 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.100155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.100189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.100198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.100216 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.100225 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.119895 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.132994 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.139497 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.140114 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.140207 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.143491 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb" exitCode=0 Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.143541 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.145020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.158373 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.170782 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.171016 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.171907 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.181478 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.192209 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.203937 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.204364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.204404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.204419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.204436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.204447 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.216798 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.230692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.242012 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.251704 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.268793 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.281241 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.293961 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.306715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.306745 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.306753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.306765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.306774 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.311237 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.344799 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.357859 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.370395 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.387224 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.399826 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.408600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.408629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.408638 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.408664 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.408674 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.413869 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.429786 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.443834 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.458057 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.477404 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.491012 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.504580 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.511402 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.511449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.511461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.511477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.511489 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.519547 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.530216 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.540400 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.551432 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.562541 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.578448 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.588444 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.601570 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.614529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.614573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.614586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.614603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.614614 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.622248 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.636390 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.649881 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.662344 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.678325 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.691435 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.706088 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.724046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.724092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.724105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.724127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.724141 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.726363 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.739572 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.751834 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.761729 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.784672 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.799317 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.826813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.826853 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.826863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.826877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.826887 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.928491 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.928530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.928542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.928559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.928570 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:40Z","lastTransitionTime":"2025-10-06T08:39:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:40 crc kubenswrapper[4989]: I1006 08:39:40.934855 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:40 crc kubenswrapper[4989]: E1006 08:39:40.934941 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.031085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.031144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.031159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.031184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.031200 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.133469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.133513 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.133524 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.133539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.133547 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.148164 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920" exitCode=0 Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.148282 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.148312 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.163788 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.176993 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.189204 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.207241 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.233447 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.237048 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.237105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.237124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.237146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.237161 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.246936 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.261729 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.275052 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.287793 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.305714 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.317029 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.327979 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.339542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.339593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.339606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.339623 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.339633 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.341029 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.359295 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.370807 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:41Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.441862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.441913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.441925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.441943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.441954 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.545577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.545692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.545715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.545742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.545762 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.649269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.649334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.649369 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.649396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.649414 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.752834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.752903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.752926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.752954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.752975 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.856322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.856408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.856432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.856462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.856483 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.935362 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.935362 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:41 crc kubenswrapper[4989]: E1006 08:39:41.935697 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:41 crc kubenswrapper[4989]: E1006 08:39:41.935565 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.958913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.958955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.958965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.958981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:41 crc kubenswrapper[4989]: I1006 08:39:41.958993 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:41Z","lastTransitionTime":"2025-10-06T08:39:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.061296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.061330 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.061346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.061363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.061374 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.154294 4989 generic.go:334] "Generic (PLEG): container finished" podID="e941df51-dd85-47d8-bb33-ab46f2c13e3c" containerID="5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6" exitCode=0 Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.154395 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerDied","Data":"5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.154437 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.163602 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.163682 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.163702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.163724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.163741 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.174087 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.188585 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.203297 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.221669 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.239597 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.255458 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.270869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.270905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.270914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.270928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.270937 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.276598 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.289700 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.312508 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.329431 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.347284 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.373609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.373669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.373681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.373699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.373711 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.384518 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.402830 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.422575 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.435758 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:42Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.475529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.475573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.475584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.475612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.475627 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.577725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.577763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.577775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.577795 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.577809 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.681232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.681260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.681268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.681282 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.681291 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.783490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.783521 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.783529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.783542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.783551 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.886228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.886265 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.886277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.886293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.886303 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.934987 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:42 crc kubenswrapper[4989]: E1006 08:39:42.935174 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.989025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.989084 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.989102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.989124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:42 crc kubenswrapper[4989]: I1006 08:39:42.989142 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:42Z","lastTransitionTime":"2025-10-06T08:39:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.091281 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.091319 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.091332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.091348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.091356 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.164046 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/0.log" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.166896 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d" exitCode=1 Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.166981 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.168393 4989 scope.go:117] "RemoveContainer" containerID="66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.170585 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" event={"ID":"e941df51-dd85-47d8-bb33-ab46f2c13e3c","Type":"ContainerStarted","Data":"cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.183184 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.193649 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.193763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.193786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.193817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.193842 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.197837 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.219977 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.235884 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.252385 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.268297 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.280766 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.295480 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.299151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.299197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.299211 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.299234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.299247 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.310343 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.326861 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.352078 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.366114 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.380062 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.390283 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.400991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.401161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.401218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.401283 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.401340 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.405626 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.418464 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.435017 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.447415 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.459183 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.468837 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.518522 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.520485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.520881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.520965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.521040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.521120 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.540944 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.552302 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.569010 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.591384 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.615469 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.623326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.623570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.623681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.623759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.623841 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.638071 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.656881 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.670824 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.685322 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:43Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.726987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.727170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.727260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.727353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.727429 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.829193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.829717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.829798 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.829884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.829961 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.931969 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.932027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.932045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.932071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.932088 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:43Z","lastTransitionTime":"2025-10-06T08:39:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.935266 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:43 crc kubenswrapper[4989]: I1006 08:39:43.935313 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:43 crc kubenswrapper[4989]: E1006 08:39:43.935419 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:43 crc kubenswrapper[4989]: E1006 08:39:43.935529 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.034877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.034921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.034933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.034951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.034964 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.137738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.137807 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.137819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.137836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.137849 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.176902 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/1.log" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.178935 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/0.log" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.183434 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b" exitCode=1 Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.183500 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.183570 4989 scope.go:117] "RemoveContainer" containerID="66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.184975 4989 scope.go:117] "RemoveContainer" containerID="73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b" Oct 06 08:39:44 crc kubenswrapper[4989]: E1006 08:39:44.185243 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.207362 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.233209 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.240064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.240099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.240110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.240127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.240139 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.249007 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.269640 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.289194 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.322242 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.343386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.343728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.343904 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.344039 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.344152 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.346504 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.364137 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.383062 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.405754 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.423700 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.442277 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.447155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.447221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.447241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.447267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.447286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.459481 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.481629 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.508924 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:44Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.550936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.550990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.551009 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.551032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.551050 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.655523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.655605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.655629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.655696 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.655759 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.758911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.758991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.759007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.759031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.759048 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.862286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.862344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.862362 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.862385 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.862402 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.935970 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:44 crc kubenswrapper[4989]: E1006 08:39:44.936453 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.965874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.965948 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.965965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.965989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:44 crc kubenswrapper[4989]: I1006 08:39:44.966004 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:44Z","lastTransitionTime":"2025-10-06T08:39:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.068042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.068105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.068173 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.068206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.068228 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.171139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.171193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.171212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.171238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.171257 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.189609 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/1.log" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.274378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.274440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.274465 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.274499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.274522 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.378128 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.378197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.378222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.378261 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.378285 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.480787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.480850 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.480871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.480896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.480913 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.583773 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.584118 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.584423 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.584822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.585180 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.611082 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.611256 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.611308 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.611355 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.611390 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611400 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611409 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611483 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:01.611458848 +0000 UTC m=+52.401484458 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611510 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:01.611498079 +0000 UTC m=+52.401523699 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611572 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611597 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611615 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611688 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611754 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611784 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611714 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:01.611696034 +0000 UTC m=+52.401721654 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.611959 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:01.611903539 +0000 UTC m=+52.401929159 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.612006 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:40:01.611988251 +0000 UTC m=+52.402013861 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.687847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.687897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.687908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.687925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.687936 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.791111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.791164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.791184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.791208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.791225 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.894255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.894345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.894364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.894387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.894406 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.898113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.898161 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.898179 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.898198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.898213 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.920429 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:45Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.925962 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.926035 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.926057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.926087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.926110 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.935748 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.935756 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.935964 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.936031 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.951363 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:45Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.954875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.955032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.955137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.955225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.955295 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.969980 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:45Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.974152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.974215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.974232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.974257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:45 crc kubenswrapper[4989]: I1006 08:39:45.974275 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:45Z","lastTransitionTime":"2025-10-06T08:39:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:45 crc kubenswrapper[4989]: E1006 08:39:45.997715 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:45Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.003246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.003380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.003439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.003497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.003566 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: E1006 08:39:46.017408 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:46Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:46 crc kubenswrapper[4989]: E1006 08:39:46.017817 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.019395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.019527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.019604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.019715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.019815 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.123115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.123493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.123717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.123912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.124119 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.228067 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.228127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.228141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.228166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.228182 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.331719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.331774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.331787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.331804 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.331816 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.434210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.434265 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.434282 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.434304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.434320 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.537047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.537390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.537512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.537579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.537632 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.640835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.641210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.641372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.641585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.641786 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.744788 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.744858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.744881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.744910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.744932 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.847864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.848129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.848204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.848584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.848689 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.934993 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:46 crc kubenswrapper[4989]: E1006 08:39:46.935192 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.954000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.954053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.954061 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.954077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:46 crc kubenswrapper[4989]: I1006 08:39:46.954087 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:46Z","lastTransitionTime":"2025-10-06T08:39:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.057296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.057605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.057787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.057963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.058108 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.161334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.161401 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.161423 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.161451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.161475 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.264359 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.264408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.264424 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.264449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.264466 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.309131 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9"] Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.310349 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.312523 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.313012 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.332970 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.344675 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.354433 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.366704 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.366738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.366746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.366759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.366768 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.367007 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.385134 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.397976 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.412379 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.422715 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.432580 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b10fe885-fdf7-4fc6-afea-9320595b680b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.432645 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqw6c\" (UniqueName: \"kubernetes.io/projected/b10fe885-fdf7-4fc6-afea-9320595b680b-kube-api-access-dqw6c\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.432774 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.432805 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.434096 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.448242 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.459751 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.469417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.469464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.469476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.469494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.469508 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.481518 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.498380 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.510963 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.522015 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.532225 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.533531 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.533641 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.533753 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b10fe885-fdf7-4fc6-afea-9320595b680b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.533852 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqw6c\" (UniqueName: \"kubernetes.io/projected/b10fe885-fdf7-4fc6-afea-9320595b680b-kube-api-access-dqw6c\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.534249 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.534276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b10fe885-fdf7-4fc6-afea-9320595b680b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.539742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b10fe885-fdf7-4fc6-afea-9320595b680b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.551487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqw6c\" (UniqueName: \"kubernetes.io/projected/b10fe885-fdf7-4fc6-afea-9320595b680b-kube-api-access-dqw6c\") pod \"ovnkube-control-plane-749d76644c-bmfk9\" (UID: \"b10fe885-fdf7-4fc6-afea-9320595b680b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.571300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.571339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.571353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.571370 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.571382 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.633709 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" Oct 06 08:39:47 crc kubenswrapper[4989]: W1006 08:39:47.646990 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb10fe885_fdf7_4fc6_afea_9320595b680b.slice/crio-3ae5dedb8ae6edbaec4407ca92b11e68f2c8c3ab84498c19f7c62d70121604f9 WatchSource:0}: Error finding container 3ae5dedb8ae6edbaec4407ca92b11e68f2c8c3ab84498c19f7c62d70121604f9: Status 404 returned error can't find the container with id 3ae5dedb8ae6edbaec4407ca92b11e68f2c8c3ab84498c19f7c62d70121604f9 Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.674372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.674403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.674411 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.674450 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.674459 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.778787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.778849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.778868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.778891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.778907 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.881476 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.881526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.881538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.881555 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.881567 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.935109 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:47 crc kubenswrapper[4989]: E1006 08:39:47.935241 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.935112 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:47 crc kubenswrapper[4989]: E1006 08:39:47.935382 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.985124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.985180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.985198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.985272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:47 crc kubenswrapper[4989]: I1006 08:39:47.985290 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:47Z","lastTransitionTime":"2025-10-06T08:39:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.088360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.088418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.088435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.088457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.088474 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.191267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.191293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.191300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.191313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.191320 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.203736 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" event={"ID":"b10fe885-fdf7-4fc6-afea-9320595b680b","Type":"ContainerStarted","Data":"7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.203777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" event={"ID":"b10fe885-fdf7-4fc6-afea-9320595b680b","Type":"ContainerStarted","Data":"a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.203788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" event={"ID":"b10fe885-fdf7-4fc6-afea-9320595b680b","Type":"ContainerStarted","Data":"3ae5dedb8ae6edbaec4407ca92b11e68f2c8c3ab84498c19f7c62d70121604f9"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.216761 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.228346 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.239871 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.250181 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.262984 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.282885 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.292967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.292998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.293007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.293020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.293032 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.297017 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.306632 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.320188 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.347438 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.358850 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.369545 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.381814 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.392788 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.395701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.395750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.395769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.395792 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.395829 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.406883 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.424286 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.497924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.497975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.497989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.498008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.498036 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.600432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.600504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.600525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.600550 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.600567 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.703705 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.704056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.704074 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.704098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.704114 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.784559 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8ndcb"] Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.785202 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:48 crc kubenswrapper[4989]: E1006 08:39:48.785289 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.806183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.806616 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.806876 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.807122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.807326 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.809810 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.827450 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.847307 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.865928 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.891574 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.910449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.910520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.910533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.910580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.910594 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:48Z","lastTransitionTime":"2025-10-06T08:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.913840 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.934411 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.934980 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:48 crc kubenswrapper[4989]: E1006 08:39:48.935246 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.946690 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrt6l\" (UniqueName: \"kubernetes.io/projected/c9cca673-1746-4fd9-bc67-330c4d06522e-kube-api-access-mrt6l\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.946757 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.947885 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.959425 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.976001 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:48 crc kubenswrapper[4989]: I1006 08:39:48.988769 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:48Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.003923 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.012718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.012765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.012782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.012806 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.012823 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.033439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.047563 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrt6l\" (UniqueName: \"kubernetes.io/projected/c9cca673-1746-4fd9-bc67-330c4d06522e-kube-api-access-mrt6l\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.047842 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.048032 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.048167 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:39:49.548131667 +0000 UTC m=+40.338157307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.053600 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.068288 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.072576 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrt6l\" (UniqueName: \"kubernetes.io/projected/c9cca673-1746-4fd9-bc67-330c4d06522e-kube-api-access-mrt6l\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.080123 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.094758 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.115417 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.115480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.115505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.115537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.115561 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.218398 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.218470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.218482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.218501 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.218515 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.321922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.321982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.322003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.322034 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.322053 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.425465 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.425559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.425576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.425598 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.425615 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.527849 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.527912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.527925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.527943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.527957 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.553365 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.553499 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.553576 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:39:50.553558296 +0000 UTC m=+41.343583876 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.630506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.630545 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.630554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.630568 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.630581 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.733365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.733439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.733462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.733490 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.733516 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.836230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.836270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.836280 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.836296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.836305 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.935050 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.935124 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.935310 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.935360 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.935513 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:49 crc kubenswrapper[4989]: E1006 08:39:49.935603 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.939428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.939483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.939503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.939536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.939561 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:49Z","lastTransitionTime":"2025-10-06T08:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.956327 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.976859 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:49 crc kubenswrapper[4989]: I1006 08:39:49.989685 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.012873 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.034713 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66440d99c968d77c5751d0ba2098ed1dfc3b1f9348ee635356745878b3b1086d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"message\\\":\\\"s:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1006 08:39:42.711309 6229 services_controller.go:452] Built service openshift-machine-api/control-plane-machine-set-operator per-node LB for network=default: []services.LB{}\\\\nI1006 08:39:42.711294 6229 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver-operator/metrics]} name:Service_openshift-kube-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.109:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {1f62a432-33b9-495d-83b2-d1dbe6961325}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1006 08:39:42.711335 6229 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.042296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.042363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.042375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.042394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.042407 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.048259 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.060463 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.070092 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.081088 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.093333 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.104206 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.126965 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.142832 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.145692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.145764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.145789 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.145820 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.145844 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.158058 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.170904 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.188480 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.200238 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.252141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.252192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.252206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.252227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.252241 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.362404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.362574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.362590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.362606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.362618 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.465783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.465863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.465887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.465919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.465945 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.563702 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:50 crc kubenswrapper[4989]: E1006 08:39:50.563852 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:50 crc kubenswrapper[4989]: E1006 08:39:50.563914 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:39:52.563897159 +0000 UTC m=+43.353922749 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.569338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.569407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.569429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.569459 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.569485 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.671773 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.671825 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.671840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.671860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.671875 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.774058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.774104 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.774119 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.774139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.774154 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.876610 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.876723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.876737 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.876756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.876769 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.935645 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:50 crc kubenswrapper[4989]: E1006 08:39:50.935877 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.979742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.979792 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.979803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.979821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:50 crc kubenswrapper[4989]: I1006 08:39:50.979835 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:50Z","lastTransitionTime":"2025-10-06T08:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.082381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.082419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.082429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.082444 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.082455 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.184983 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.185050 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.185066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.185088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.185112 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.288372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.288464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.288496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.288527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.288548 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.391865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.391917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.391936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.391959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.391976 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.495103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.495169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.495193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.495222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.495244 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.598445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.598508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.598527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.598554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.598574 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.702054 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.702115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.702131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.702153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.702171 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.805210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.805296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.805313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.805358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.805371 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.908911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.908952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.908964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.908981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.908993 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:51Z","lastTransitionTime":"2025-10-06T08:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.935618 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:51 crc kubenswrapper[4989]: E1006 08:39:51.935801 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.935944 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:51 crc kubenswrapper[4989]: I1006 08:39:51.935970 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:51 crc kubenswrapper[4989]: E1006 08:39:51.936088 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:51 crc kubenswrapper[4989]: E1006 08:39:51.936190 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.011736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.011769 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.011780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.011794 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.011806 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.114800 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.114863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.114893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.114922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.114945 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.217542 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.217813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.217828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.217844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.217856 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.320448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.320474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.320483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.320497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.320507 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.424094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.424152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.424164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.424190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.424205 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.527068 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.527137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.527164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.527197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.527222 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.588303 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:52 crc kubenswrapper[4989]: E1006 08:39:52.588501 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:52 crc kubenswrapper[4989]: E1006 08:39:52.588601 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:39:56.588576368 +0000 UTC m=+47.378601988 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.630958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.631045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.631073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.631097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.631119 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.733389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.733427 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.733435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.733451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.733462 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.835389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.835447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.835460 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.835477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.835489 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.934937 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:52 crc kubenswrapper[4989]: E1006 08:39:52.935073 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.938080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.938160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.938190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.938230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:52 crc kubenswrapper[4989]: I1006 08:39:52.938259 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:52Z","lastTransitionTime":"2025-10-06T08:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.041190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.041241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.041256 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.041277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.041290 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.143166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.143204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.143221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.143242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.143253 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.192499 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.193981 4989 scope.go:117] "RemoveContainer" containerID="73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b" Oct 06 08:39:53 crc kubenswrapper[4989]: E1006 08:39:53.194342 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.212757 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.231631 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.245381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.245416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.245427 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.245442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.245453 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.248250 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.265987 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.276990 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.288601 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.299575 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.311109 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.324271 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.338047 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.347896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.347929 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.347940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.347956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.347971 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.358300 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.376906 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.388458 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.400149 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.411498 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.425813 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.436625 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:53Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.450303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.450328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.450339 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.450354 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.450364 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.553323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.553372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.553386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.553403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.553414 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.656158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.656205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.656222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.656242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.656257 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.758940 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.759006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.759018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.759035 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.759049 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.861776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.861881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.861905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.861939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.861963 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.935768 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.935802 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.935802 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:53 crc kubenswrapper[4989]: E1006 08:39:53.935941 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:53 crc kubenswrapper[4989]: E1006 08:39:53.936150 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:53 crc kubenswrapper[4989]: E1006 08:39:53.936292 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.963641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.963701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.963712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.963727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:53 crc kubenswrapper[4989]: I1006 08:39:53.963737 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:53Z","lastTransitionTime":"2025-10-06T08:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.066748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.066813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.066830 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.066856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.066875 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.170156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.170199 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.170213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.170239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.170252 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.273247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.273288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.273297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.273313 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.273324 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.376779 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.376858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.376876 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.376899 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.376916 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.480164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.480247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.480259 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.480276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.480288 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.583053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.583119 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.583133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.583148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.583161 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.689335 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.689505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.689527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.689566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.689584 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.792379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.792434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.792451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.792472 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.792488 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.895731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.895794 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.895812 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.895835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.895854 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.935089 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:54 crc kubenswrapper[4989]: E1006 08:39:54.935463 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.998765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.998867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.998885 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.998911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:54 crc kubenswrapper[4989]: I1006 08:39:54.998928 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:54Z","lastTransitionTime":"2025-10-06T08:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.102425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.102503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.102528 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.102558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.102583 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.205053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.205428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.205608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.205914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.206099 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.310024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.310089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.310107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.310131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.310148 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.411958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.412197 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.412255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.412318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.412377 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.514987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.515039 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.515085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.515107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.515124 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.618089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.618143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.618164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.618189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.618207 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.721037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.721107 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.721136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.721168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.721196 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.824609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.824673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.824685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.824702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.824714 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.927874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.927954 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.927980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.928010 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.928033 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:55Z","lastTransitionTime":"2025-10-06T08:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.935871 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.935984 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:55 crc kubenswrapper[4989]: I1006 08:39:55.936104 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:55 crc kubenswrapper[4989]: E1006 08:39:55.936138 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:55 crc kubenswrapper[4989]: E1006 08:39:55.936073 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:55 crc kubenswrapper[4989]: E1006 08:39:55.936250 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.030149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.030585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.030793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.030952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.031105 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.133589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.133628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.133649 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.133687 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.133700 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.236125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.236167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.236178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.236195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.236208 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.295909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.295982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.296006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.296039 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.296061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.312614 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:56Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.317232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.317372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.317455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.317536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.317609 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.329500 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:56Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.333982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.334064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.334089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.334122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.334156 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.350361 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:56Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.354260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.354482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.354560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.354680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.354760 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.365914 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:56Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.369365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.369416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.369432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.369455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.369475 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.380977 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:56Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.381137 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.382743 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.382775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.382784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.382799 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.382811 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.485536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.485575 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.485584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.485599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.485609 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.588255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.588813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.588887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.588965 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.589032 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.630301 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.630833 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.631104 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:40:04.631073514 +0000 UTC m=+55.421099134 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.692279 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.692572 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.692695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.692797 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.692874 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.796317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.796387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.796411 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.796439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.796461 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.899469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.899525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.899534 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.899558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.899568 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:56Z","lastTransitionTime":"2025-10-06T08:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:56 crc kubenswrapper[4989]: I1006 08:39:56.935537 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:56 crc kubenswrapper[4989]: E1006 08:39:56.935738 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.002972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.003367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.003525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.003709 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.003897 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.107602 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.107644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.107686 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.107704 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.107716 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.210264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.210296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.210306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.210320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.210330 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.216222 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.229220 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.239089 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.261629 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.283499 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.299158 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.312859 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.313333 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.313393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.313411 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.313437 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.313455 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.336221 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.354715 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.371899 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.390737 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.414399 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.416294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.416331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.416346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.416367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.416380 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.430844 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.448808 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.461432 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.481529 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.499489 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.516145 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.519529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.519589 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.519615 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.519641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.519688 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.534812 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:57Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.622992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.623081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.623108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.623133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.623152 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.727140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.727223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.727235 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.727258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.727275 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.830034 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.830062 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.830070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.830083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.830092 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.932797 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.932950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.932964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.932988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.933007 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:57Z","lastTransitionTime":"2025-10-06T08:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.935245 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.935343 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:57 crc kubenswrapper[4989]: I1006 08:39:57.935343 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:57 crc kubenswrapper[4989]: E1006 08:39:57.935500 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:57 crc kubenswrapper[4989]: E1006 08:39:57.935630 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:57 crc kubenswrapper[4989]: E1006 08:39:57.935892 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.035348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.035399 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.035412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.035430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.035445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.138171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.138253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.138266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.138285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.138297 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.240770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.240821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.240833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.240862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.240876 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.343348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.343435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.343461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.343487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.343505 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.446402 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.446444 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.446453 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.446468 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.446479 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.549870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.549939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.549956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.549980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.549997 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.652430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.652486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.652496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.652513 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.652525 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.755895 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.755969 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.755985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.756002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.756013 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.859435 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.859486 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.859497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.859513 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.859525 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.935816 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:39:58 crc kubenswrapper[4989]: E1006 08:39:58.936002 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.961710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.961752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.961761 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.961774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:58 crc kubenswrapper[4989]: I1006 08:39:58.961784 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:58Z","lastTransitionTime":"2025-10-06T08:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.065052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.065106 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.065118 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.065136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.065150 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.168320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.168379 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.168393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.168415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.168429 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.271088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.271136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.271154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.271175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.271190 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.374409 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.374477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.374497 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.374525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.374542 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.481215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.481406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.481421 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.481436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.481448 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.583957 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.584054 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.584112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.584137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.584154 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.686538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.686596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.686614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.686637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.686689 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.790008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.790040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.790048 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.790063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.790073 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.891973 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.892017 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.892028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.892044 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.892055 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.935517 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.935525 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:39:59 crc kubenswrapper[4989]: E1006 08:39:59.935640 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:39:59 crc kubenswrapper[4989]: E1006 08:39:59.935728 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.936497 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:39:59 crc kubenswrapper[4989]: E1006 08:39:59.936638 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.949082 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:59Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.966765 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:59Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.979125 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:59Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.992262 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:39:59Z is after 2025-08-24T17:21:41Z" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.994362 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.994392 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.994403 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.994422 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:39:59 crc kubenswrapper[4989]: I1006 08:39:59.994435 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:39:59Z","lastTransitionTime":"2025-10-06T08:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.003751 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.021863 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.034572 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.042615 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.053414 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.072086 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.083003 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.093007 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.096345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.096462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.096533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.096596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.096672 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.104277 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.114618 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.129213 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.139874 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.152014 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.164579 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:00Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.200363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.200408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.200426 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.200449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.200468 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.302957 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.303015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.303033 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.303057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.303073 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.406528 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.406598 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.406620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.406644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.406693 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.509353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.509383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.509392 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.509406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.509419 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.612612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.612924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.613028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.613098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.613153 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.715891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.716440 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.716519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.716590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.716650 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.819140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.819173 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.819184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.819202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.819215 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.921752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.922006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.922121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.922227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.922308 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:00Z","lastTransitionTime":"2025-10-06T08:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:00 crc kubenswrapper[4989]: I1006 08:40:00.935049 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:00 crc kubenswrapper[4989]: E1006 08:40:00.935135 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.025634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.025729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.025747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.025774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.025792 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.127844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.127886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.127895 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.127908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.127918 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.230448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.230510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.230526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.230549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.230566 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.333442 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.333510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.333532 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.333562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.333583 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.435793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.435821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.435831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.435844 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.435853 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.538285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.538317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.538328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.538344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.538361 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.641166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.641212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.641224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.641240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.641250 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.695159 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.695306 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.695345 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.695387 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.695424 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695505 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:40:33.69546204 +0000 UTC m=+84.485487620 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695534 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695616 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:33.695606364 +0000 UTC m=+84.485631934 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695619 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695710 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695733 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695747 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695804 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:33.695786628 +0000 UTC m=+84.485812218 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695713 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695830 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695860 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:33.69585254 +0000 UTC m=+84.485878130 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.695926 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.696046 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:40:33.696011924 +0000 UTC m=+84.486037504 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.744817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.744855 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.744864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.744882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.744892 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.846799 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.846837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.846848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.846863 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.846875 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.935182 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.935215 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.935259 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.935357 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.935475 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:01 crc kubenswrapper[4989]: E1006 08:40:01.935606 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.948966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.949113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.949196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.949278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:01 crc kubenswrapper[4989]: I1006 08:40:01.949366 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:01Z","lastTransitionTime":"2025-10-06T08:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.051971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.052027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.052050 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.052070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.052084 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.154457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.154877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.155013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.155149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.155292 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.257794 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.257859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.257871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.257888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.257899 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.360523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.360561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.360570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.360583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.360593 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.463085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.463131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.463142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.463160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.463172 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.566432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.566480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.566492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.566514 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.566528 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.668735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.668813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.668829 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.668851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.668867 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.771056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.771668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.771750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.771816 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.771905 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.874803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.874852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.874865 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.874881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.874892 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.935310 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:02 crc kubenswrapper[4989]: E1006 08:40:02.935442 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.977786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.977869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.977887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.977910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:02 crc kubenswrapper[4989]: I1006 08:40:02.977927 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:02Z","lastTransitionTime":"2025-10-06T08:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.080819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.080878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.080898 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.080926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.080946 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.184300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.184706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.184888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.185066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.185238 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.287811 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.287856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.287867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.287884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.287896 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.391007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.391400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.391532 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.391737 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.391887 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.495162 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.495208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.495219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.495237 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.495268 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.597540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.597599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.597611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.597634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.597650 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.700192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.700249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.700266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.700290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.700310 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.803194 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.803253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.803271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.803298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.803318 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.905994 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.906044 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.906063 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.906083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.906093 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:03Z","lastTransitionTime":"2025-10-06T08:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.935824 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.935854 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:03 crc kubenswrapper[4989]: E1006 08:40:03.935938 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:03 crc kubenswrapper[4989]: I1006 08:40:03.935997 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:03 crc kubenswrapper[4989]: E1006 08:40:03.936021 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:03 crc kubenswrapper[4989]: E1006 08:40:03.936124 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.008924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.008969 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.008980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.008998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.009009 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.111018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.111058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.111068 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.111085 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.111097 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.213195 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.213278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.213294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.213314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.213329 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.316426 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.316505 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.316525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.316556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.316580 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.419682 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.419721 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.419793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.419811 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.419822 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.522253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.522293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.522304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.522318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.522328 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.625201 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.625261 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.625271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.625284 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.625294 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.728019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.728071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.728091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.728114 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.728132 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.729459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:04 crc kubenswrapper[4989]: E1006 08:40:04.729608 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:04 crc kubenswrapper[4989]: E1006 08:40:04.729723 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:40:20.729695205 +0000 UTC m=+71.519720845 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.830625 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.830750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.830780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.830815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.830841 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.934907 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:04 crc kubenswrapper[4989]: E1006 08:40:04.935061 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.935316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.935340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.935348 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.935392 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:04 crc kubenswrapper[4989]: I1006 08:40:04.935401 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:04Z","lastTransitionTime":"2025-10-06T08:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.037966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.038015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.038032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.038053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.038068 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.140609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.140673 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.140685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.140700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.140711 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.243091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.243190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.243206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.243225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.243236 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.347091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.347135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.347148 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.347171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.347187 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.449934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.450002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.450020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.450045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.450061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.552731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.553072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.553088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.553112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.553124 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.655413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.655452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.655461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.655474 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.655485 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.758174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.758231 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.758239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.758254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.758263 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.861499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.861537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.861547 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.861564 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.861574 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.935074 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.935124 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:05 crc kubenswrapper[4989]: E1006 08:40:05.935211 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.935290 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:05 crc kubenswrapper[4989]: E1006 08:40:05.935418 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:05 crc kubenswrapper[4989]: E1006 08:40:05.935470 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.964991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.965041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.965053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.965071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:05 crc kubenswrapper[4989]: I1006 08:40:05.965088 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:05Z","lastTransitionTime":"2025-10-06T08:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.068680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.068744 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.068759 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.068784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.068805 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.172917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.172983 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.172995 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.173015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.173029 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.275939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.275976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.275986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.276002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.276012 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.378879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.378934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.378945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.378963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.378975 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.419512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.419552 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.419563 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.419575 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.419584 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.433909 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:06Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.444503 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.444557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.444569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.444588 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.444601 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.460036 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:06Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.464938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.464990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.465003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.465022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.465035 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.480180 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:06Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.485153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.485228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.485240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.485270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.485300 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.501394 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:06Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.506631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.506694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.506706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.506727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.506739 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.520209 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:06Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.520349 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.522172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.522202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.522243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.522258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.522267 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.624876 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.624964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.624975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.624991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.625017 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.727543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.727614 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.727642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.727688 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.727702 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.831258 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.831298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.831307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.831321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.831331 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.932953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.932993 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.933009 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.933092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.933128 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:06Z","lastTransitionTime":"2025-10-06T08:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:06 crc kubenswrapper[4989]: I1006 08:40:06.935221 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:06 crc kubenswrapper[4989]: E1006 08:40:06.935447 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.035103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.035136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.035145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.035160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.035169 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.137896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.137951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.137963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.137985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.137998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.240858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.240918 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.240935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.240961 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.240978 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.343599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.343681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.343720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.343751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.343773 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.447489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.447543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.447554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.447570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.447580 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.550880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.550934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.550951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.550974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.550992 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.654114 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.654155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.654169 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.654185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.654197 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.758003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.758049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.758060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.758083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.758094 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.861703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.861783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.861803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.861832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.861850 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.935875 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.935903 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.935927 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:07 crc kubenswrapper[4989]: E1006 08:40:07.936628 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:07 crc kubenswrapper[4989]: E1006 08:40:07.936838 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:07 crc kubenswrapper[4989]: E1006 08:40:07.936943 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.937252 4989 scope.go:117] "RemoveContainer" containerID="73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.965144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.965208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.965234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.965263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:07 crc kubenswrapper[4989]: I1006 08:40:07.965286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:07Z","lastTransitionTime":"2025-10-06T08:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.067835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.067900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.067923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.067947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.067965 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.170373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.170410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.170421 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.170438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.170449 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.273523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.273582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.273604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.273633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.273676 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.277458 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/1.log" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.280900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.282457 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.301954 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.317154 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.332140 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.371124 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.375601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.375633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.375643 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.375672 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.375681 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.393225 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.413272 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.429643 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.448234 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.465926 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.477125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.477156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.477164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.477176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.477185 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.480043 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.495196 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.511773 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.535055 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.547035 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.567780 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579340 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579733 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579743 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.579765 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.590925 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.600240 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.683391 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.683432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.683445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.683462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.683473 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.786101 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.786146 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.786156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.786173 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.786184 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.888693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.888731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.888741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.888758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.888768 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.935626 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:08 crc kubenswrapper[4989]: E1006 08:40:08.935796 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.991332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.991374 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.991386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.991406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:08 crc kubenswrapper[4989]: I1006 08:40:08.991420 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:08Z","lastTransitionTime":"2025-10-06T08:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.093879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.093914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.093925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.093941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.093949 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.195910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.195947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.195958 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.195972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.195984 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.284535 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/2.log" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.285442 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/1.log" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.288181 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" exitCode=1 Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.288213 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.288249 4989 scope.go:117] "RemoveContainer" containerID="73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.290682 4989 scope.go:117] "RemoveContainer" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" Oct 06 08:40:09 crc kubenswrapper[4989]: E1006 08:40:09.291281 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.297496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.297543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.297553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.297568 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.297577 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.313892 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.327453 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.340414 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.363757 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.379600 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.393511 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.399791 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.399828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.399840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.399856 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.399868 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.409581 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.424804 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.442745 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.458929 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.478723 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.490568 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.502483 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.502529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.502539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.502557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.502566 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.506020 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.533788 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.549481 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.566119 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.579914 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.595198 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.605296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.605337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.605349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.605367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.605378 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.708266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.708334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.708358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.708386 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.708409 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.812055 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.812119 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.812138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.812163 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.812181 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.915812 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.915851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.915882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.915897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.915907 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:09Z","lastTransitionTime":"2025-10-06T08:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.934922 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:09 crc kubenswrapper[4989]: E1006 08:40:09.935051 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.935885 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:09 crc kubenswrapper[4989]: E1006 08:40:09.936153 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.936349 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:09 crc kubenswrapper[4989]: E1006 08:40:09.936430 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.951003 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.972324 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:09 crc kubenswrapper[4989]: I1006 08:40:09.992690 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:09Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.012638 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.017877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.017915 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.017928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.017947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.017963 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.027995 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.042107 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.056151 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.090343 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.111034 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.120089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.120125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.120137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.120155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.120166 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.123560 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.143056 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.164934 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73bc916fb105f914b68dd114718f3e7c7b502e21345093b6d905d241e7a7299b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:39:44Z\\\",\\\"message\\\":\\\"alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000068 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1006 08:39:44.000112 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000148 6416 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000182 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000302 6416 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1006 08:39:44.000798 6416 factory.go:656] Stopping watch factory\\\\nI1006 08:39:44.009597 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1006 08:39:44.009620 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1006 08:39:44.009676 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1006 08:39:44.009694 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1006 08:39:44.009752 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.185965 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.199824 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.209771 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.220461 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.222177 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.222204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.222213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.222228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.222238 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.232047 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.244015 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.293626 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/2.log" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.303448 4989 scope.go:117] "RemoveContainer" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" Oct 06 08:40:10 crc kubenswrapper[4989]: E1006 08:40:10.303631 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.317488 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.324648 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.324753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.324777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.324803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.324813 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.328198 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.348445 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.363310 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.376315 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.391014 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.405391 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.422076 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.427113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.427149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.427160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.427176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.427186 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.437691 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.452699 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.467451 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.481382 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.496672 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.518044 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.530296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.530336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.530347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.530367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.530378 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.531604 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.543339 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.551941 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.565765 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:10Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.633071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.633121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.633141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.633167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.633185 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.736244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.736307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.736324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.736351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.736372 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.838718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.838790 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.838816 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.838848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.838872 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.935096 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:10 crc kubenswrapper[4989]: E1006 08:40:10.935261 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.940838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.940930 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.940955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.940984 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:10 crc kubenswrapper[4989]: I1006 08:40:10.941003 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:10Z","lastTransitionTime":"2025-10-06T08:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.044955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.044997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.045012 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.045030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.045041 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.150232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.150268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.150276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.150290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.150299 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.253944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.253989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.254000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.254020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.254033 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.357176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.357232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.357249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.357273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.357290 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.460015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.460098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.460110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.460128 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.460138 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.562895 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.562972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.562989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.563025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.563061 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.666140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.666175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.666185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.666200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.666208 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.769541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.769597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.769609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.769628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.769641 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.872091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.872144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.872157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.872175 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.872187 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.935176 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.935271 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:11 crc kubenswrapper[4989]: E1006 08:40:11.935420 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.935472 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:11 crc kubenswrapper[4989]: E1006 08:40:11.935643 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:11 crc kubenswrapper[4989]: E1006 08:40:11.935850 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.975000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.975051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.975068 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.975090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:11 crc kubenswrapper[4989]: I1006 08:40:11.975106 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:11Z","lastTransitionTime":"2025-10-06T08:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.077543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.077574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.077582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.077597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.077605 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.179799 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.179884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.179905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.179931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.179952 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.283908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.283979 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.284002 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.284029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.284051 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.387042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.387131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.387153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.387178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.387197 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.490334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.490395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.490412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.490438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.490455 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.593816 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.593885 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.593907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.593939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.593960 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.697429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.697491 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.697557 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.697582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.697597 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.799860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.799896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.799904 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.799916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.799927 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.902494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.902556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.902579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.902608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.902630 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:12Z","lastTransitionTime":"2025-10-06T08:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:12 crc kubenswrapper[4989]: I1006 08:40:12.935743 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:12 crc kubenswrapper[4989]: E1006 08:40:12.936240 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.005192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.005250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.005266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.005288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.005304 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.107506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.107543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.107554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.107569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.107579 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.210608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.210700 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.210723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.210753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.210776 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.312455 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.312520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.312538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.312560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.312577 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.414919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.415016 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.415042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.415755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.415998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.518288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.518323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.518331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.518346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.518356 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.621716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.621754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.621768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.621786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.621797 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.724831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.724906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.724925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.724949 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.724966 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.827500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.827567 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.827579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.827604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.827619 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.930031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.930070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.930086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.930104 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.930118 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:13Z","lastTransitionTime":"2025-10-06T08:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.935594 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.935688 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:13 crc kubenswrapper[4989]: E1006 08:40:13.935736 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:13 crc kubenswrapper[4989]: I1006 08:40:13.935787 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:13 crc kubenswrapper[4989]: E1006 08:40:13.935912 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:13 crc kubenswrapper[4989]: E1006 08:40:13.935953 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.032364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.032402 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.032413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.032432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.032445 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.135328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.135377 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.135387 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.135400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.135415 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.238453 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.238492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.238504 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.238522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.238533 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.340760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.340811 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.340829 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.340871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.340888 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.443162 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.443204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.443219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.443240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.443255 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.546546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.546596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.546608 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.546631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.546643 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.649247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.649288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.649300 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.649316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.649329 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.751200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.751242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.751252 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.751269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.751281 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.854141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.854204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.854215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.854230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.854239 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.934929 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:14 crc kubenswrapper[4989]: E1006 08:40:14.935069 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.957180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.957223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.957232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.957247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:14 crc kubenswrapper[4989]: I1006 08:40:14.957276 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:14Z","lastTransitionTime":"2025-10-06T08:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.059382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.059432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.059445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.059464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.059479 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.161559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.161632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.161642 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.161669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.161680 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.263905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.263943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.263953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.263966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.263974 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.365839 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.365868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.365878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.365890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.365900 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.468144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.468176 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.468187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.468204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.468216 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.570835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.570869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.570880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.570896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.570908 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.672907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.672947 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.672961 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.672981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.672992 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.775744 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.775785 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.775794 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.775808 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.775818 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.877576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.877625 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.877634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.877680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.877691 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.935304 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.935367 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:15 crc kubenswrapper[4989]: E1006 08:40:15.935421 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.935304 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:15 crc kubenswrapper[4989]: E1006 08:40:15.935503 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:15 crc kubenswrapper[4989]: E1006 08:40:15.935751 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.980358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.980549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.980618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.980739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:15 crc kubenswrapper[4989]: I1006 08:40:15.980821 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:15Z","lastTransitionTime":"2025-10-06T08:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.083481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.083535 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.083551 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.083577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.083593 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.186249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.186283 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.186294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.186311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.186321 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.288903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.289168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.289250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.289316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.289407 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.391963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.392004 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.392012 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.392026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.392036 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.494174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.494605 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.494835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.495083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.495378 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.597478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.597520 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.597528 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.597544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.597553 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.655431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.656036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.656160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.656293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.656399 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.670414 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:16Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.677263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.677308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.677331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.677352 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.677367 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.691010 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:16Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.694775 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.694822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.694837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.694858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.694872 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.707120 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:16Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.710681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.710725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.710736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.710752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.710763 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.721501 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:16Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.725302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.725342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.725352 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.725364 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.725374 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.739851 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:16Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.740356 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.742259 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.742277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.742285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.742312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.742321 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.844803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.844892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.844909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.844961 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.844978 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.935542 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:16 crc kubenswrapper[4989]: E1006 08:40:16.935780 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.947127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.947353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.947412 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.947469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:16 crc kubenswrapper[4989]: I1006 08:40:16.947527 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:16Z","lastTransitionTime":"2025-10-06T08:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.050512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.050801 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.050921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.051007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.051082 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.153456 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.153514 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.153541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.153562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.153580 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.256689 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.257317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.257419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.257514 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.257603 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.359698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.359945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.360066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.360164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.360248 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.463080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.463129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.463141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.463159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.463172 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.565485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.565802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.565935 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.566029 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.566098 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.668230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.668290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.668299 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.668312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.668320 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.770584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.770624 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.770634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.770669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.770689 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.873046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.873108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.873128 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.873152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.873169 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.935934 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:17 crc kubenswrapper[4989]: E1006 08:40:17.936165 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.936347 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:17 crc kubenswrapper[4989]: E1006 08:40:17.936578 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.936615 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:17 crc kubenswrapper[4989]: E1006 08:40:17.937100 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.949532 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.975793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.975858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.975878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.975901 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:17 crc kubenswrapper[4989]: I1006 08:40:17.975918 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:17Z","lastTransitionTime":"2025-10-06T08:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.079143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.079204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.079221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.079243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.079258 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.181081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.181133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.181141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.181155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.181164 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.283649 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.283708 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.283720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.283735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.283746 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.386692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.386731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.386739 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.386752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.386762 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.489582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.489626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.489637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.489668 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.489680 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.592389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.592437 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.592448 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.592466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.592478 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.694843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.694872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.694880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.694892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.694901 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.797089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.797139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.797152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.797170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.797182 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.900088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.900172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.900193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.900222 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.900245 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:18Z","lastTransitionTime":"2025-10-06T08:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:18 crc kubenswrapper[4989]: I1006 08:40:18.935609 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:18 crc kubenswrapper[4989]: E1006 08:40:18.935778 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.002315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.002372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.002388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.002410 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.002427 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.104554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.104595 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.104604 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.104621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.104632 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.207073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.207113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.207126 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.207144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.207158 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.309229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.309266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.309275 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.309290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.309300 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.411735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.411797 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.411814 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.411837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.411855 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.513924 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.513963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.513972 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.513986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.513998 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.616584 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.616706 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.616735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.616767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.616791 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.719358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.719407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.719416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.719431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.719440 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.822267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.822316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.822327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.822346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.822358 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.924204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.924278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.924298 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.924324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.924344 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:19Z","lastTransitionTime":"2025-10-06T08:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.935498 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.935532 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.935532 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:19 crc kubenswrapper[4989]: E1006 08:40:19.935675 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:19 crc kubenswrapper[4989]: E1006 08:40:19.935767 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:19 crc kubenswrapper[4989]: E1006 08:40:19.935859 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.950402 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:19Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.961114 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:19Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.971736 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:19Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.984903 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:19Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:19 crc kubenswrapper[4989]: I1006 08:40:19.994923 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:19Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.012102 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.022137 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.031167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.031212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.031224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.031242 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.031253 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.032631 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.043571 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.056663 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.077505 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.089487 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.101333 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.118174 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.133525 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.133846 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.134202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.134217 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.134238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.134250 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.145796 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.155545 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.164505 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.175470 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:20Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.236752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.236778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.236787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.236800 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.236809 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.339334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.339394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.339404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.339419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.339431 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.442124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.442191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.442209 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.442238 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.442280 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.544481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.544669 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.544763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.544838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.544900 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.647906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.647960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.647977 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.647999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.648013 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.750165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.750219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.750237 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.750263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.750280 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.804159 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:20 crc kubenswrapper[4989]: E1006 08:40:20.804329 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:20 crc kubenswrapper[4989]: E1006 08:40:20.804444 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:40:52.804418532 +0000 UTC m=+103.594444152 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.853489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.853715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.853808 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.853911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.853987 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.936031 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:20 crc kubenswrapper[4989]: E1006 08:40:20.936260 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.956620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.956762 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.956864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.956966 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:20 crc kubenswrapper[4989]: I1006 08:40:20.957058 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:20Z","lastTransitionTime":"2025-10-06T08:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.060019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.060082 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.060100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.060129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.060150 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.162794 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.162840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.162852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.162872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.162884 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.266289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.266331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.266340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.266355 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.266364 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.368980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.369053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.369070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.369094 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.369112 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.472288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.472359 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.472382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.472416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.472437 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.576124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.576163 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.576180 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.576203 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.576216 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.678817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.678890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.678913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.678944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.678968 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.782110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.782189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.782219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.782251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.782275 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.885578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.885694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.885719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.885747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.885766 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.935282 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:21 crc kubenswrapper[4989]: E1006 08:40:21.935497 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.935614 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:21 crc kubenswrapper[4989]: E1006 08:40:21.935883 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.935967 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:21 crc kubenswrapper[4989]: E1006 08:40:21.936065 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.988093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.988156 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.988174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.988199 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:21 crc kubenswrapper[4989]: I1006 08:40:21.988217 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:21Z","lastTransitionTime":"2025-10-06T08:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.090718 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.090784 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.090801 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.090824 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.090841 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.193597 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.193641 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.193666 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.193683 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.193695 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.296645 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.296725 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.296735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.296751 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.296760 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.399758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.399815 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.399833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.399857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.399876 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.503322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.503388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.503404 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.503432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.503450 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.606269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.606316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.606328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.606344 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.606356 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.708917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.708970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.708984 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.709006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.709019 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.811627 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.811681 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.811695 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.811713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.811722 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.915033 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.915098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.915111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.915129 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.915140 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:22Z","lastTransitionTime":"2025-10-06T08:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:22 crc kubenswrapper[4989]: I1006 08:40:22.934912 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:22 crc kubenswrapper[4989]: E1006 08:40:22.935133 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.017462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.017561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.017577 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.017593 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.017606 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.120052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.120091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.120103 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.120120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.120130 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.222869 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.222907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.222916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.222946 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.222974 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.325405 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.325452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.325464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.325481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.325494 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.343338 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/0.log" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.343408 4989 generic.go:334] "Generic (PLEG): container finished" podID="b901f1ca-c342-4faa-a715-0ff3d94bc369" containerID="9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69" exitCode=1 Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.343448 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerDied","Data":"9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.343949 4989 scope.go:117] "RemoveContainer" containerID="9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.361973 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.374268 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.385912 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.405727 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.425881 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.428174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.428210 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.428219 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.428234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.428244 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.438783 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.451561 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.465695 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.479758 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.491354 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.502726 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.512781 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.530206 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.546311 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.546998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.547349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.547373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.547393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.547406 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.567751 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.581837 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.593517 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.604230 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.617778 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:23Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.649692 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.649738 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.649749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.649767 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.649780 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.751988 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.752039 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.752055 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.752077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.752094 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.854078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.854125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.854138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.854155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.854167 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.935471 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.935547 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:23 crc kubenswrapper[4989]: E1006 08:40:23.935646 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.935701 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:23 crc kubenswrapper[4989]: E1006 08:40:23.936092 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:23 crc kubenswrapper[4989]: E1006 08:40:23.936202 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.936546 4989 scope.go:117] "RemoveContainer" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" Oct 06 08:40:23 crc kubenswrapper[4989]: E1006 08:40:23.936786 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.956910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.956980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.957006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.957037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:23 crc kubenswrapper[4989]: I1006 08:40:23.957058 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:23Z","lastTransitionTime":"2025-10-06T08:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.060528 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.060600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.060623 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.060698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.060725 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.164046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.164110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.164121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.164138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.164151 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.268302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.268342 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.268353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.268368 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.268377 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.349500 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/0.log" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.349575 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerStarted","Data":"2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.366021 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.371633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.371734 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.371758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.371789 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.371811 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.383516 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.399010 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.417316 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.442439 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.461809 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.475452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.475502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.475519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.475545 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.475562 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.478594 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.491421 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.508820 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.545384 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.566560 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.578489 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.578537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.578548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.578562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.578574 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.581112 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.602558 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.623071 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.650484 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.668237 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.682466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.682512 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.682523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.682540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.682557 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.685536 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.704000 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.720848 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:24Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.784993 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.785030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.785043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.785057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.785067 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.888236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.888288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.888302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.888320 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.888334 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.935734 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:24 crc kubenswrapper[4989]: E1006 08:40:24.935866 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.990756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.990792 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.990803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.990818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:24 crc kubenswrapper[4989]: I1006 08:40:24.990829 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:24Z","lastTransitionTime":"2025-10-06T08:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.094360 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.094389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.094400 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.094416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.094427 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.196943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.197008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.197020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.197036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.197051 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.300093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.300163 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.300183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.300205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.300223 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.403874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.403926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.403943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.403967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.403984 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.506847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.506883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.506896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.506913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.506924 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.609897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.609939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.609950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.609967 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.609978 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.712434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.712485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.712496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.712513 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.712525 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.816720 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.816785 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.816803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.816828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.816846 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.920558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.920633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.920691 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.920724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.920761 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:25Z","lastTransitionTime":"2025-10-06T08:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.935409 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.935466 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:25 crc kubenswrapper[4989]: E1006 08:40:25.935645 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:25 crc kubenswrapper[4989]: I1006 08:40:25.935764 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:25 crc kubenswrapper[4989]: E1006 08:40:25.935910 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:25 crc kubenswrapper[4989]: E1006 08:40:25.936076 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.023044 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.023120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.023133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.023153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.023165 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.126032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.126065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.126075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.126089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.126098 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.228964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.229031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.229055 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.229086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.229111 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.332625 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.332719 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.332742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.332771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.332795 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.435652 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.435832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.435850 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.435878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.435899 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.538763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.538848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.538882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.538912 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.538937 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.642184 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.642223 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.642233 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.642251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.642263 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.745905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.746007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.746077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.746105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.746210 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.849172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.849224 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.849244 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.849269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.849287 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.935649 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:26 crc kubenswrapper[4989]: E1006 08:40:26.935910 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.952363 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.952418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.952434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.952457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:26 crc kubenswrapper[4989]: I1006 08:40:26.952475 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:26Z","lastTransitionTime":"2025-10-06T08:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.016297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.016349 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.016371 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.016396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.016417 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.037847 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:27Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.042685 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.042746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.042765 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.042796 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.042816 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.062269 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:27Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.066920 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.066955 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.066970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.066989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.067003 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.084340 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:27Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.088817 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.088866 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.088884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.088909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.088926 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.103128 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:27Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.106888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.106934 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.106951 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.106974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.106992 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.121363 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:27Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.121511 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.122724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.122745 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.122756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.122770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.122780 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.225747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.225820 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.225842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.225873 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.225895 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.329192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.329253 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.329277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.329307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.329331 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.432494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.432576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.432601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.432631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.432696 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.535220 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.535270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.535289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.535312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.535330 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.638937 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.639028 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.639047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.639099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.639117 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.741838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.741883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.741892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.741915 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.741929 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.845603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.845670 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.845682 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.845699 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.845711 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.935495 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.935584 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.935717 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.935723 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.935782 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:27 crc kubenswrapper[4989]: E1006 08:40:27.936014 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.948049 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.948136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.948152 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.948174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:27 crc kubenswrapper[4989]: I1006 08:40:27.948191 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:27Z","lastTransitionTime":"2025-10-06T08:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.054185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.054232 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.054243 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.054259 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.054270 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.157780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.157827 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.157842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.157860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.157873 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.260637 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.260730 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.260748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.260770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.260787 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.362882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.362933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.362945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.362962 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.362975 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.465714 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.465789 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.465802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.465823 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.465838 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.569165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.569245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.569269 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.569299 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.569324 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.671914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.671986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.672004 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.672027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.672044 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.775519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.775594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.775753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.775803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.775827 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.879458 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.879511 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.879522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.879539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.879548 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.935171 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:28 crc kubenswrapper[4989]: E1006 08:40:28.935310 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.982111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.982151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.982165 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.982182 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:28 crc kubenswrapper[4989]: I1006 08:40:28.982194 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:28Z","lastTransitionTime":"2025-10-06T08:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.091007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.091061 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.091099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.091121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.091135 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.195485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.195625 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.195643 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.195709 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.195736 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.298080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.298123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.298135 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.298151 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.298163 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.400834 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.400936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.400959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.400990 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.401013 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.503851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.503916 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.503936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.503959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.503976 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.606818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.606879 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.606892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.606911 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.606923 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.709684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.709732 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.709798 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.709822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.709837 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.813378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.813449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.813467 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.813492 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.813508 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.916473 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.916539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.916558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.916578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.916593 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:29Z","lastTransitionTime":"2025-10-06T08:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.935170 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.935256 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:29 crc kubenswrapper[4989]: E1006 08:40:29.935376 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.935402 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:29 crc kubenswrapper[4989]: E1006 08:40:29.935579 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:29 crc kubenswrapper[4989]: E1006 08:40:29.935719 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.952974 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:29Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.969106 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:29Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.982955 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:29Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:29 crc kubenswrapper[4989]: I1006 08:40:29.996788 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:29Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.010628 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.018832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.018867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.018880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.018897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.018910 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.030809 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.045717 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.077756 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.096387 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.113842 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.121297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.121336 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.121347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.121361 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.121372 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.132213 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.153482 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.174502 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.192710 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.211402 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.224929 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.225008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.225032 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.225064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.225086 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.233132 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.252303 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.275761 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.305296 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:30Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.327598 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.327648 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.327680 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.327698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.327711 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.430073 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.430111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.430127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.430147 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.430161 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.533228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.533288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.533304 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.533326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.533342 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.635777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.636052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.636087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.636115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.636135 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.739852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.739928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.739953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.739987 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.740012 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.843199 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.843317 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.843340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.843372 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.843391 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.935687 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:30 crc kubenswrapper[4989]: E1006 08:40:30.935844 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.945644 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.945783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.945804 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.945827 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:30 crc kubenswrapper[4989]: I1006 08:40:30.945843 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:30Z","lastTransitionTime":"2025-10-06T08:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.049723 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.049838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.049862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.049891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.049915 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.153027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.153093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.153112 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.153142 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.153164 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.256058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.256144 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.256170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.256200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.256226 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.359133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.359215 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.359225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.359245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.359256 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.462562 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.462640 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.462712 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.462746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.462779 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.565621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.565703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.565717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.565736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.565748 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.668585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.668708 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.668735 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.668766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.668792 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.771901 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.771963 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.771981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.772006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.772024 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.875696 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.875772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.875799 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.875832 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.875851 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.935443 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.935551 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.935557 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:31 crc kubenswrapper[4989]: E1006 08:40:31.935709 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:31 crc kubenswrapper[4989]: E1006 08:40:31.935884 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:31 crc kubenswrapper[4989]: E1006 08:40:31.936015 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.979396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.979457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.979478 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.979500 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:31 crc kubenswrapper[4989]: I1006 08:40:31.979517 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:31Z","lastTransitionTime":"2025-10-06T08:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.082553 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.082617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.082635 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.082711 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.082740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.186192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.186251 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.186267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.186290 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.186307 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.288978 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.289051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.289079 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.289111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.289134 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.391783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.391828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.391837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.391852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.391864 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.495188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.495236 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.495250 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.495268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.495282 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.597850 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.597936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.597959 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.597992 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.598014 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.700717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.700788 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.700831 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.700864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.700886 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.803327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.803390 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.803408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.803432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.803448 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.906962 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.907020 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.907040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.907065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.907083 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:32Z","lastTransitionTime":"2025-10-06T08:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:32 crc kubenswrapper[4989]: I1006 08:40:32.934982 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:32 crc kubenswrapper[4989]: E1006 08:40:32.935166 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.009843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.009907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.009919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.009956 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.009968 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.113635 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.113860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.113891 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.113922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.113947 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.217031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.217105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.217123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.217145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.217162 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.320522 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.320591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.320613 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.320649 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.320697 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.424071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.424147 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.424171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.424203 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.424228 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.527595 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.527646 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.527742 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.527766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.527836 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.630571 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.630698 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.630727 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.630758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.630783 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.733202 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.733262 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.733274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.733294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.733310 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.746179 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.746293 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.746366 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746387 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:37.746360733 +0000 UTC m=+148.536386323 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746457 4989 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746518 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:41:37.746502007 +0000 UTC m=+148.536527597 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746517 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746551 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.746459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746570 4989 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746638 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746797 4989 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746816 4989 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746822 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-06 08:41:37.746781444 +0000 UTC m=+148.536807085 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.746802 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.746893 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-06 08:41:37.746877237 +0000 UTC m=+148.536902907 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.747054 4989 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.747206 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-06 08:41:37.747137824 +0000 UTC m=+148.537163414 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.836903 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.836964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.836981 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.837007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.837025 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.935764 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.935822 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.935941 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.935976 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.936141 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:33 crc kubenswrapper[4989]: E1006 08:40:33.936277 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.939827 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.939857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.939868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.939884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:33 crc kubenswrapper[4989]: I1006 08:40:33.939895 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:33Z","lastTransitionTime":"2025-10-06T08:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.042270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.042330 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.042345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.042365 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.042379 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.145702 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.145741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.145753 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.145771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.145783 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.248141 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.248239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.248263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.248293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.248315 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.352477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.352572 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.352591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.352612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.352629 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.455496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.455560 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.455574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.455596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.455622 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.557894 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.558001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.558026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.558051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.558071 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.661334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.661389 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.661405 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.661428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.661455 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.764451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.764508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.764529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.764552 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.764569 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.867230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.867296 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.867306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.867329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.867342 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.935813 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:34 crc kubenswrapper[4989]: E1006 08:40:34.935944 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.970643 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.970722 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.970731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.970750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:34 crc kubenswrapper[4989]: I1006 08:40:34.970762 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:34Z","lastTransitionTime":"2025-10-06T08:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.073648 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.073724 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.073734 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.073757 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.073770 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.177454 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.177515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.177548 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.177586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.177608 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.280177 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.280263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.280286 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.280314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.280338 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.383350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.383419 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.383436 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.383462 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.383479 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.486383 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.486463 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.486487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.486539 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.486562 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.588802 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.588881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.588906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.588931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.588948 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.692240 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.692294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.692310 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.692331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.692350 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.794711 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.794764 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.794774 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.794787 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.794796 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.898323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.898391 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.898406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.898425 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.898441 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:35Z","lastTransitionTime":"2025-10-06T08:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.935243 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.935269 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:35 crc kubenswrapper[4989]: E1006 08:40:35.935450 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:35 crc kubenswrapper[4989]: I1006 08:40:35.935572 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:35 crc kubenswrapper[4989]: E1006 08:40:35.935751 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:35 crc kubenswrapper[4989]: E1006 08:40:35.935856 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.000939 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.001018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.001042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.001072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.001096 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.104388 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.104499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.104523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.104555 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.104580 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.207138 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.207168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.207178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.207190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.207198 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.309884 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.309922 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.309933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.309948 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.309958 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.412164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.412212 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.412227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.412247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.412259 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.515904 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.515964 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.515982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.516007 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.516030 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.618982 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.619024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.619038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.619056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.619071 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.721621 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.721701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.721721 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.721746 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.721763 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.824544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.824583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.824594 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.824611 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.824622 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.927185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.927254 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.927271 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.927294 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.927311 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:36Z","lastTransitionTime":"2025-10-06T08:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.935595 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:36 crc kubenswrapper[4989]: E1006 08:40:36.935824 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:36 crc kubenswrapper[4989]: I1006 08:40:36.936965 4989 scope.go:117] "RemoveContainer" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.031334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.031429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.031447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.031488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.031504 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.134057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.134311 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.134322 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.134338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.134348 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.236853 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.236902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.236913 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.236928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.236939 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.344323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.344373 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.344395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.344413 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.344426 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.407068 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/2.log" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.409588 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.410552 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.426636 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.446270 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.446302 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.446312 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.446327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.446338 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.447019 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.457854 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.457999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.458027 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.458037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.458051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.458062 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.468917 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.472906 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.476087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.476128 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.476139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.476157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.476167 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.479066 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.489428 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.493235 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.493277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.493288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.493308 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.493320 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.494322 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.504895 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.507120 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.510711 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.510748 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.510757 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.510772 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.510783 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.516742 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.520252 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.522580 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.522616 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.522626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.522640 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.522661 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.529421 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.532759 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.532916 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.541329 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.548530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.548558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.548565 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.548579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.548588 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.551607 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.573198 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.584542 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.594375 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.608041 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.626865 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.638501 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.650850 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.650880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.650890 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.650905 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.650916 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.651021 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.659750 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.753040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.753093 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.753108 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.753127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.753138 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.856493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.856538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.856551 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.856569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.856585 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.935468 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.935554 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.935508 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.935736 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.935934 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:37 crc kubenswrapper[4989]: E1006 08:40:37.936021 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.959617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.959731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.959763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.959793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:37 crc kubenswrapper[4989]: I1006 08:40:37.959817 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:37Z","lastTransitionTime":"2025-10-06T08:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.062353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.062393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.062406 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.062434 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.062456 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.169177 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.169241 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.169257 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.169278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.169293 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.272943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.273014 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.273031 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.273056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.273073 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.376876 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.376928 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.376945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.376971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.376989 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.417105 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/3.log" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.418319 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/2.log" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.423697 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" exitCode=1 Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.423764 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.423829 4989 scope.go:117] "RemoveContainer" containerID="1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.429400 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:40:38 crc kubenswrapper[4989]: E1006 08:40:38.429884 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.449175 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.469571 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.480105 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.480168 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.480188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.480211 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.480228 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.487316 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.504985 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.523217 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.539318 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.571287 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.583917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.583985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.584008 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.584038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.584145 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.589171 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.605747 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.654258 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.686536 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.686570 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.686578 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.686591 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.686601 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.688785 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1209653f030489e13754e92794c007c4784c3bec0ad7594913c5e72e0ed575d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:08Z\\\",\\\"message\\\":\\\"for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:08Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:08.909897 6707 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-authentication/oauth-openshift]} name:Service_openshift-authentication/oauth-openshift_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.222:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c0c2f725-e461-454e-a88c-c8350d62e1ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1006 08:40:08.909905 6707 services_controller.go:443] Built service openshift-\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"er-node configs for network=default: []services.lbConfig(nil)\\\\nI1006 08:40:37.807237 7103 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1006 08:40:37.807255 7103 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:37.807257 7103 services_controller.go:451] Built service openshift-authentication-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-authentication-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", E\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.701198 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.713547 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.726552 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.735414 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.745290 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.757720 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.765647 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.774697 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:38Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.789218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.789255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.789263 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.789276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.789286 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.893003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.893057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.893076 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.893099 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.893116 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.934990 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:38 crc kubenswrapper[4989]: E1006 08:40:38.935194 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.995607 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.995689 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.995713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.995740 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:38 crc kubenswrapper[4989]: I1006 08:40:38.995757 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:38Z","lastTransitionTime":"2025-10-06T08:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.098650 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.098736 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.098754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.098777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.098795 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.201480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.201521 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.201533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.201550 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.201561 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.304120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.304190 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.304207 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.304234 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.304251 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.407754 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.407818 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.407837 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.407868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.407886 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.431011 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/3.log" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.435809 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:40:39 crc kubenswrapper[4989]: E1006 08:40:39.436056 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.457495 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.489084 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"er-node configs for network=default: []services.lbConfig(nil)\\\\nI1006 08:40:37.807237 7103 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1006 08:40:37.807255 7103 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:37.807257 7103 services_controller.go:451] Built service openshift-authentication-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-authentication-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", E\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.507824 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.513558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.513600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.513613 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.513635 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.513668 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.531235 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.544477 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.560230 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.577576 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.596242 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.613079 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.616407 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.616461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.616485 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.616517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.616539 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.628355 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.649147 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.665475 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.680341 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.698533 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.716584 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.718498 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.718543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.718555 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.718574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.718586 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.727898 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.758725 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.775634 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.793757 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.821880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.821933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.821957 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.821984 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.822005 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.925071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.925139 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.925164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.925198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.925224 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:39Z","lastTransitionTime":"2025-10-06T08:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.935174 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:39 crc kubenswrapper[4989]: E1006 08:40:39.935393 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.935774 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:39 crc kubenswrapper[4989]: E1006 08:40:39.935906 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.936015 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:39 crc kubenswrapper[4989]: E1006 08:40:39.936163 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.957676 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.972454 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:39 crc kubenswrapper[4989]: I1006 08:40:39.985927 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.000692 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:39Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.016535 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.028285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.028341 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.028358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.028381 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.028398 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.033368 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.047408 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.061357 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.072965 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.098089 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.113200 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.131185 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.131214 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.131225 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.131239 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.131252 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.132142 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.151758 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.165346 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.184422 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.201497 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.214867 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.232010 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.233354 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.233375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.233382 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.233394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.233403 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.256511 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"er-node configs for network=default: []services.lbConfig(nil)\\\\nI1006 08:40:37.807237 7103 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1006 08:40:37.807255 7103 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:37.807257 7103 services_controller.go:451] Built service openshift-authentication-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-authentication-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", E\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:40Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.336015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.336075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.336092 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.336115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.336132 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.438600 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.439047 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.439186 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.439318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.439466 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.542882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.542936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.542953 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.542976 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.542994 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.649583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.649628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.649645 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.649715 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.649740 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.752131 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.752428 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.752508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.752571 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.752633 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.856072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.856123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.856160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.856188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.856209 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.935081 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:40 crc kubenswrapper[4989]: E1006 08:40:40.935273 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.959496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.959561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.959579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.959603 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:40 crc kubenswrapper[4989]: I1006 08:40:40.959620 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:40Z","lastTransitionTime":"2025-10-06T08:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.062888 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.063914 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.063978 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.064013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.064039 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.165887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.165937 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.165950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.165970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.165982 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.268378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.268433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.268445 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.268463 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.268480 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.370974 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.371050 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.371059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.371072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.371081 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.473797 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.473848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.473859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.473877 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.473890 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.577022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.577070 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.577081 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.577098 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.577115 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.679525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.679590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.679607 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.679630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.679641 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.783021 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.783059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.783068 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.783080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.783088 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.885065 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.885110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.885121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.885137 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.885155 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.935800 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.935915 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:41 crc kubenswrapper[4989]: E1006 08:40:41.935947 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:41 crc kubenswrapper[4989]: E1006 08:40:41.936073 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.936089 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:41 crc kubenswrapper[4989]: E1006 08:40:41.936524 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.987632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.987716 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.987763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.987792 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:41 crc kubenswrapper[4989]: I1006 08:40:41.987824 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:41Z","lastTransitionTime":"2025-10-06T08:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.091100 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.091150 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.091160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.091178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.091189 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.195208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.195277 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.195299 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.195327 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.195348 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.301932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.301983 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.301999 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.302025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.302043 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.405927 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.406026 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.406038 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.406055 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.406066 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.509297 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.509346 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.509359 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.509376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.509388 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.612838 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.612906 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.612925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.612950 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.612967 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.716978 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.717022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.717035 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.717052 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.717064 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.820752 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.820803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.820813 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.820828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.820837 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.923897 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.923925 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.923932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.923944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.923985 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:42Z","lastTransitionTime":"2025-10-06T08:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:42 crc kubenswrapper[4989]: I1006 08:40:42.935894 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:42 crc kubenswrapper[4989]: E1006 08:40:42.936161 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.026524 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.026559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.026569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.026586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.026596 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.129089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.129158 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.129171 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.129188 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.129203 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.231874 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.231931 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.231943 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.231961 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.231973 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.334907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.334998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.335023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.335051 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.335078 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.437882 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.437944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.437960 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.437986 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.438004 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.541731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.541776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.541785 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.541798 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.541808 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.645072 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.645260 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.645292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.645375 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.645400 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.748115 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.748160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.748173 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.748192 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.748204 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.850585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.850694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.850717 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.850747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.850771 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.935941 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.936004 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.936405 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:43 crc kubenswrapper[4989]: E1006 08:40:43.936639 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:43 crc kubenswrapper[4989]: E1006 08:40:43.937003 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:43 crc kubenswrapper[4989]: E1006 08:40:43.937156 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.953285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.953337 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.953353 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.953376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:43 crc kubenswrapper[4989]: I1006 08:40:43.953393 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:43Z","lastTransitionTime":"2025-10-06T08:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.055690 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.055766 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.055777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.055793 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.055804 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.159496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.159556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.159574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.159601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.159618 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.262064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.262110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.262121 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.262136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.262149 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.365036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.365087 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.365101 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.365125 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.365141 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.467750 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.467826 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.467842 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.467862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.467877 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.571255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.571511 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.571556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.571586 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.571609 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.674230 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.674289 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.674306 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.674331 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.674348 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.777821 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.777900 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.777919 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.777944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.777962 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.881227 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.881280 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.881292 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.881310 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.881358 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.934974 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:44 crc kubenswrapper[4989]: E1006 08:40:44.935141 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.983952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.983985 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.984001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.984019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:44 crc kubenswrapper[4989]: I1006 08:40:44.984030 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:44Z","lastTransitionTime":"2025-10-06T08:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.086731 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.086768 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.086778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.086822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.086838 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.188587 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.188622 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.188630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.188645 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.188671 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.291540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.291583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.291596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.291618 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.291629 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.393612 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.393674 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.393686 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.393703 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.393715 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.496268 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.496321 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.496340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.496362 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.496379 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.599796 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.599871 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.599894 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.599923 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.599945 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.702278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.702324 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.702335 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.702350 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.702360 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.806083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.806155 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.806170 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.806194 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.806208 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.909421 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.909494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.909518 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.909546 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.909623 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:45Z","lastTransitionTime":"2025-10-06T08:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.935010 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.935100 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:45 crc kubenswrapper[4989]: I1006 08:40:45.935022 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:45 crc kubenswrapper[4989]: E1006 08:40:45.935245 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:45 crc kubenswrapper[4989]: E1006 08:40:45.935389 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:45 crc kubenswrapper[4989]: E1006 08:40:45.935467 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.011332 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.011380 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.011394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.011415 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.011429 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.114102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.114177 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.114200 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.114228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.114251 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.217022 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.217090 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.217109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.217130 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.217151 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.319780 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.319847 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.319867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.319893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.319910 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.422726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.422809 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.422828 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.422851 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.422869 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.526123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.526187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.526205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.526229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.526257 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.631037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.631113 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.631133 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.631157 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.631182 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.734043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.734083 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.734091 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.734109 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.734121 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.836468 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.836523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.836538 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.836556 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.836569 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.935538 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:46 crc kubenswrapper[4989]: E1006 08:40:46.935682 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.939860 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.939883 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.939892 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.939902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:46 crc kubenswrapper[4989]: I1006 08:40:46.939912 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:46Z","lastTransitionTime":"2025-10-06T08:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.042908 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.043178 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.043247 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.043314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.043372 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.147015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.147329 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.147523 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.147707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.147864 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.250590 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.250635 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.250647 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.250684 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.250698 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.353439 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.353506 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.353526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.353551 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.353568 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.457431 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.457466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.457475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.457488 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.457497 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.560001 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.560046 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.560058 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.560075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.560094 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.662569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.662686 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.662707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.662729 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.662747 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.766416 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.766487 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.766511 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.766541 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.766889 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.870003 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.870033 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.870042 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.870054 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.870062 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.897881 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.897917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.897926 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.897941 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.897950 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.916145 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.920835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.920886 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.920901 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.920921 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.920934 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.935162 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.935253 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.935487 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.935619 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.935838 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.935936 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.939970 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.944858 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.944896 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.944904 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.944920 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.944931 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.961143 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.965378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.965422 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.965433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.965452 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.965496 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.979928 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.984206 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.984411 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.984558 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.984760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:47 crc kubenswrapper[4989]: I1006 08:40:47.984929 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:47Z","lastTransitionTime":"2025-10-06T08:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.999684 4989 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"98580f71-44d1-4b79-83d9-4c80b9cd1cfc\\\",\\\"systemUUID\\\":\\\"ef74357a-4028-4330-ae84-d5f17f42c6f5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:47Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:47 crc kubenswrapper[4989]: E1006 08:40:47.999914 4989 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.001870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.001932 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.001948 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.001970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.001986 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.104707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.104763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.104778 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.104795 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.104809 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.207288 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.207326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.207334 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.207347 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.207357 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.311276 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.311352 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.311376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.311405 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.311452 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.415318 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.415395 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.415418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.415447 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.415470 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.519356 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.519393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.519408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.519438 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.519455 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.622132 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.622189 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.622205 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.622229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.622247 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.725422 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.725475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.725493 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.725519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.725538 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.828726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.828797 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.828820 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.828848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.828869 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.931582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.931694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.931749 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.931781 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.931798 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:48Z","lastTransitionTime":"2025-10-06T08:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:48 crc kubenswrapper[4989]: I1006 08:40:48.936016 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:48 crc kubenswrapper[4989]: E1006 08:40:48.936232 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.035110 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.035160 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.035174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.035191 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.035204 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.138508 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.138551 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.138559 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.138574 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.138582 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.241968 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.242024 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.242039 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.242059 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.242073 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.344756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.344833 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.344857 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.344885 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.344907 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.448314 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.448376 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.448393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.448418 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.448436 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.550305 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.550430 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.550449 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.550477 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.550494 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.653429 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.653567 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.653601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.653629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.653691 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.757433 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.757579 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.757602 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.757626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.757646 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.860535 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.860620 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.860693 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.860726 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.860749 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.935291 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.935544 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.935280 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:49 crc kubenswrapper[4989]: E1006 08:40:49.935738 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:49 crc kubenswrapper[4989]: E1006 08:40:49.935897 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:49 crc kubenswrapper[4989]: E1006 08:40:49.936058 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.953694 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ddvqd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c2dc1b-134d-4ee4-af3b-8143ed4ac2f0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04659d80b70998aae5018f659f76b70a9f19fca11bdbfdec76b1d0e1b539e29b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wb6pv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ddvqd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.968630 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.968854 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.968875 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.968917 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.968937 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:49Z","lastTransitionTime":"2025-10-06T08:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:49 crc kubenswrapper[4989]: I1006 08:40:49.979568 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e941df51-dd85-47d8-bb33-ab46f2c13e3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cfe61f55de35321e5983cfb4a43abc029d3bf8b5216453461552be9e1332368a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8cc7a5c32120819ba22f700ef254ca7b8ef42c08a1fedfe97a241621b02bfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://037d237a400dbcaa0001e72d1885c8411d50b8e17db87053f4c31919329392d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0badeb7d4b0e79b6d9db1e5ca0818e5afff846aeaf8223186a78c1774d653b09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dd46de1041d140e6c02b7846e5ef9b611e5f797096df0464b6e30ade5eeccb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e747992ddb46d4046baea5420e2c03171df6d4d57d90b0f4c2086dd0fd49b920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5cd89d75fc5450a953bceed243667d0bb62fefa5450cb36e71d5cf7dd25dfba6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcqm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5s7m9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:49Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.012131 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"91d0cca1-dfc6-4f12-a25a-7325782935ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:37Z\\\",\\\"message\\\":\\\"er-node configs for network=default: []services.lbConfig(nil)\\\\nI1006 08:40:37.807237 7103 services_controller.go:445] Built service openshift-authentication-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1006 08:40:37.807255 7103 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:37Z is after 2025-08-24T17:21:41Z]\\\\nI1006 08:40:37.807257 7103 services_controller.go:451] Built service openshift-authentication-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-authentication-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", E\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:40:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nl6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-t6mdb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.031526 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.046022 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.055438 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7b8zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70b2c0ba-ddd3-41e6-ae65-5d0e9a4b3d77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb2fa7bbe51a8471b47685661ce4419ea4c687585be5d0ef4c3d06e1630835da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xcb2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:33Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7b8zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.066389 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10fe885-fdf7-4fc6-afea-9320595b680b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0488758661cbfd1aa2b191f1b0df4707c7e826c05d69cef51e3adcab862afc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7476e8d54c9f697ee4a6479f612458279c028a29218c3eab58d29e508b3199dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dqw6c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bmfk9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.071264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.071307 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.071357 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.071374 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.071386 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.079926 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cd67893-7328-49ff-8c45-d58deb3aa560\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76a3636ae59c4f5c019d0fb6ab7f35d9fbb93852b727bca600cc34d9b02df022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://557560c6ab967f4490676d95ab4056281919f2fa348e0d498800126814d1d152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc296bde0e5021426158d75886457318667ec3d659c67e1487679686b1ab677c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba416bcba21867e64209d8b84c0af172288211bd34ec0581bc088a1c422c7505\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.094738 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5108bf93a26bf9e989fd1142ca7bb137dde1c058b1f2efcbd6cd5c9920bd0127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.106620 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9cca673-1746-4fd9-bc67-330c4d06522e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrt6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8ndcb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.118214 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0450a531-c6be-4c49-bd48-cf5c405a32e9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3549861b1687ac5d38c4781e3647c7785f17084224598235bb3e6b1c26aef550\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ac52d70eff73836670ae949fe04cab2b60b3cc422fe1580fcfeb38c8de27e7c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.131324 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wz8dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b901f1ca-c342-4faa-a715-0ff3d94bc369\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:40:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-06T08:40:22Z\\\",\\\"message\\\":\\\"2025-10-06T08:39:37+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1\\\\n2025-10-06T08:39:37+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1cfb8f7e-57ba-46ab-ba80-997dca8a7ff1 to /host/opt/cni/bin/\\\\n2025-10-06T08:39:37Z [verbose] multus-daemon started\\\\n2025-10-06T08:39:37Z [verbose] Readiness Indicator file check\\\\n2025-10-06T08:40:22Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:36Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:40:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4sh5x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wz8dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.146252 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1ffbbcc8-0c75-46af-9fc7-2ac4e9680374\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8428743711cfaefe9a097cd1942efb637e0e41adba6d54eeaa890bd437f2861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45ad624ecc711ddfdbc388d441e6970bc5b4d98c500d03acbfb4bcfd3d52a97d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fcf99fc9df9eb6748551db8e19ca87a645b2eb8ca0e1cbb5ba62aec325c7a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edfda88fb2f55231dc6fc883827ef91ace6ccce47b68d0ba9b6494e064f7edb0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.163478 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.173472 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.173515 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.173526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.173544 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.173557 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.182965 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a58dd0254d152ba781078f6f6551bc34582b0a0c3ce1a30fda147ea588de32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.200349 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65c82bf276932273c9b1387fe5d478373b31c845ee9642eb4bdf2a7751edaf03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8478a9717f2d5961d84f43f6da8e7b49721501149be9cea1a46655b70d972b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.217865 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f78ef9f2-a706-4a6c-8a9f-b232f97e9211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e66241190ffcc7fb8f23936ef89b9bf4dd6214969e8cc4ec1727fd3af3973e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rz8hg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-v6vsq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.254919 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2abeb261-39e3-4712-969f-d94c1ab3cb9e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a566a9c42bd7cf5018e2bf4424d2cb78efea912603120986531f04792b17cfab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bea350cf99f692baebb0bef102c752fa7a8e8fc5a073e7b2b70315c3de52b33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759a16fb2bd0e4e6a75ebf223cd3c45a5e1b64e1b9354c28deff556a15646191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b668ef9a66da21b0fd767f2ce67cf0ffa015d14ee41e53f77e8b46b0aadf980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb50f0eade8c5d56f7a7f8462ddf98b672bf7689ee3c050d5cdc1990b9adcf21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7b6b26c5201a5f6fd08f2db646ab029ed166050de1987a6f5e951e24d346d568\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f5196ee3529c6a95f404bad8b6e4db9038a56960d1428151763cd6ed69f0576\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe2522af55819a984536beb516a8fc584b0dc4e4d9b9a390abd0d9f1b53ca64b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.270921 4989 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d856cfa9-24fa-4ce6-ac0e-b1a8066d0310\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-06T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ae889797b12777f32dc3e592c4568f40c32a725162f71a18ce17bb5c0c3815f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1205c69d313f83fa5852a5715df5cd2e3165511f87efc4064d9006ffb5b4ccbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f1aa54350eb2d59bd00dcf851276cff1d80f1c8b077ca222db41f617b95c243\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd6c2606d59c66a9c1bf5d95a87548858a8bf43955b6b974594e241e89f66440\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b034791e1a148223bf64efb5563b156d3865d494fe3fd8ba1a16cb463aac7fb0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-06T08:39:23Z\\\",\\\"message\\\":\\\"W1006 08:39:13.194601 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1006 08:39:13.195011 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759739953 cert, and key in /tmp/serving-cert-3307977662/serving-signer.crt, /tmp/serving-cert-3307977662/serving-signer.key\\\\nI1006 08:39:13.393082 1 observer_polling.go:159] Starting file observer\\\\nW1006 08:39:13.395405 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1006 08:39:13.395838 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1006 08:39:13.397672 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3307977662/tls.crt::/tmp/serving-cert-3307977662/tls.key\\\\\\\"\\\\nF1006 08:39:23.535051 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://446fb6aa58894e0bd5e8068600a5d5c58a93d44efc85d9ae3a6fd4db6830792c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-06T08:39:12Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4693fcdd8b648c4a0178361ca7967ce9cf38f1896fd6b0c8909fa54eff0999a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-06T08:39:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-06T08:39:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-06T08:39:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-06T08:40:50Z is after 2025-08-24T17:21:41Z" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.276980 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.277245 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.277357 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.277457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.277543 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.380394 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.380453 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.380469 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.380494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.380533 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.483777 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.483840 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.483862 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.483889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.483910 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.587208 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.587273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.587291 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.587315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.587333 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.690808 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.690870 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.690887 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.690910 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.690926 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.794291 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.794369 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.794392 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.794424 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.794446 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.897971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.898057 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.898088 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.898120 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.898142 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:50Z","lastTransitionTime":"2025-10-06T08:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.935944 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:50 crc kubenswrapper[4989]: E1006 08:40:50.936140 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:50 crc kubenswrapper[4989]: I1006 08:40:50.937327 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:40:50 crc kubenswrapper[4989]: E1006 08:40:50.937577 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.000852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.001097 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.001267 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.003000 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.003079 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.107221 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.107303 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.107323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.107351 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.107370 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.214326 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.214369 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.214378 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.214393 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.214402 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.318711 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.318782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.318795 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.318820 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.318835 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.422036 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.422127 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.422145 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.422172 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.422192 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.525902 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.525970 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.525989 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.526014 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.526031 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.628868 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.629249 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.629414 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.629576 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.629749 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.732701 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.732771 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.732786 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.732803 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.732814 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.835457 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.835517 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.835617 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.835745 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.835775 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.935857 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.935975 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:51 crc kubenswrapper[4989]: E1006 08:40:51.936107 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.936211 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:51 crc kubenswrapper[4989]: E1006 08:40:51.936438 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:51 crc kubenswrapper[4989]: E1006 08:40:51.936581 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.938632 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.938747 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.938760 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.938776 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:51 crc kubenswrapper[4989]: I1006 08:40:51.938789 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:51Z","lastTransitionTime":"2025-10-06T08:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.042123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.042181 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.042193 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.042213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.042227 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.145293 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.145358 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.145367 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.145391 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.145402 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.248019 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.248069 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.248078 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.248096 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.248109 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.351533 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.351609 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.351634 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.351713 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.351741 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.455273 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.455443 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.455470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.455587 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.455648 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.559016 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.559089 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.559123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.559154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.559176 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.662848 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.662938 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.662975 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.663004 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.663037 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.766864 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.766991 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.767023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.767075 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.767100 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.870196 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.870264 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.870287 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.870315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.870333 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.882846 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:52 crc kubenswrapper[4989]: E1006 08:40:52.883289 4989 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:52 crc kubenswrapper[4989]: E1006 08:40:52.883510 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs podName:c9cca673-1746-4fd9-bc67-330c4d06522e nodeName:}" failed. No retries permitted until 2025-10-06 08:41:56.883470105 +0000 UTC m=+167.673495745 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs") pod "network-metrics-daemon-8ndcb" (UID: "c9cca673-1746-4fd9-bc67-330c4d06522e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.935945 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:52 crc kubenswrapper[4989]: E1006 08:40:52.936260 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.972785 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.972854 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.972872 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.972894 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:52 crc kubenswrapper[4989]: I1006 08:40:52.972911 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:52Z","lastTransitionTime":"2025-10-06T08:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.077527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.077585 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.077601 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.077631 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.077647 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.180136 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.180187 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.180204 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.180226 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.180242 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.283502 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.283581 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.283606 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.283633 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.283703 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.386880 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.386952 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.386971 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.386994 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.387016 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.491071 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.491143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.491166 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.491198 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.491220 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.593936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.593984 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.593996 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.594013 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.594025 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.696758 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.696796 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.696806 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.696819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.696829 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.799396 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.799456 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.799468 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.799496 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.799514 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.903266 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.903316 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.903328 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.903345 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.903356 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:53Z","lastTransitionTime":"2025-10-06T08:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.935712 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.935984 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:53 crc kubenswrapper[4989]: I1006 08:40:53.936097 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:53 crc kubenswrapper[4989]: E1006 08:40:53.936278 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:53 crc kubenswrapper[4989]: E1006 08:40:53.936412 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:53 crc kubenswrapper[4989]: E1006 08:40:53.936537 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.006755 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.006835 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.006861 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.006893 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.006915 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.110174 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.110237 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.110255 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.110278 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.110295 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.213466 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.213525 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.213543 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.213573 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.213596 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.317167 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.317228 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.317246 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.317274 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.317295 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.420464 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.420516 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.420534 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.420554 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.420572 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.523756 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.523824 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.523841 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.523867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.523886 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.627045 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.627140 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.627159 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.627183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.627203 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.730408 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.730494 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.730530 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.730561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.730584 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.833790 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.833859 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.833878 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.833907 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.833923 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.934982 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:54 crc kubenswrapper[4989]: E1006 08:40:54.935187 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.938077 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.938132 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.938154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.938183 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:54 crc kubenswrapper[4989]: I1006 08:40:54.938205 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:54Z","lastTransitionTime":"2025-10-06T08:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.041475 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.041534 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.041566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.041599 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.041622 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.145569 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.145629 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.145646 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.145678 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.145726 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.248670 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.248801 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.248819 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.248843 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.248860 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.352064 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.352194 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.352218 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.352248 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.352269 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.454933 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.455006 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.455030 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.455060 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.455084 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.559338 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.559441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.559480 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.559510 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.559529 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.662441 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.662519 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.662537 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.662561 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.662579 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.766451 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.766836 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.766867 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.766889 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.766906 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.873984 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.874062 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.874086 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.874123 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.874158 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.935744 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.935822 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.935851 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:55 crc kubenswrapper[4989]: E1006 08:40:55.936041 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:55 crc kubenswrapper[4989]: E1006 08:40:55.936367 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:55 crc kubenswrapper[4989]: E1006 08:40:55.936527 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.977728 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.977800 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.977825 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.977852 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:55 crc kubenswrapper[4989]: I1006 08:40:55.977872 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:55Z","lastTransitionTime":"2025-10-06T08:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.081040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.081101 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.081124 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.081154 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.081179 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.184704 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.184770 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.184792 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.184822 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.184847 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.288588 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.288657 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.288707 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.288730 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.288747 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.392044 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.392102 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.392122 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.392149 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.392170 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.495909 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.495997 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.496025 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.496056 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.496080 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.599596 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.599649 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.599710 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.599741 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.599761 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.702426 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.702470 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.702481 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.702499 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.702510 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.805043 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.805111 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.805130 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.805153 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.805170 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.908936 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.908998 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.909015 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.909040 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.909060 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:56Z","lastTransitionTime":"2025-10-06T08:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:56 crc kubenswrapper[4989]: I1006 08:40:56.935706 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:56 crc kubenswrapper[4989]: E1006 08:40:56.935870 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.012213 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.012272 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.012291 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.012315 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.012334 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.115066 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.115126 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.115143 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.115164 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.115184 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.218461 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.218549 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.218582 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.218613 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.218637 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.322628 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.322763 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.322782 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.322804 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.322822 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.425945 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.426018 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.426037 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.426062 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.426079 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.529527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.529583 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.529602 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.529626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.529643 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.631694 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.631783 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.631801 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.631824 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.631840 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.735229 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.735285 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.735301 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.735323 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.735339 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.837944 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.838023 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.838053 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.838080 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.838102 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.936015 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.936057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.936038 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:57 crc kubenswrapper[4989]: E1006 08:40:57.936221 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:57 crc kubenswrapper[4989]: E1006 08:40:57.936380 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:57 crc kubenswrapper[4989]: E1006 08:40:57.936547 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.940526 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.940566 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.940575 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.940587 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:57 crc kubenswrapper[4989]: I1006 08:40:57.940595 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:57Z","lastTransitionTime":"2025-10-06T08:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.048041 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.048340 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.048432 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.048527 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.048611 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:58Z","lastTransitionTime":"2025-10-06T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.057482 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.057529 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.057540 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.057555 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.057567 4989 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-06T08:40:58Z","lastTransitionTime":"2025-10-06T08:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.104417 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd"] Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.105155 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.106805 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.107315 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.107543 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.107548 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.120120 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=41.120103337 podStartE2EDuration="41.120103337s" podCreationTimestamp="2025-10-06 08:40:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.119231704 +0000 UTC m=+108.909257294" watchObservedRunningTime="2025-10-06 08:40:58.120103337 +0000 UTC m=+108.910128927" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.147724 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-wz8dx" podStartSLOduration=85.147705064 podStartE2EDuration="1m25.147705064s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.1342238 +0000 UTC m=+108.924249420" watchObservedRunningTime="2025-10-06 08:40:58.147705064 +0000 UTC m=+108.937730644" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.211878 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.211862971 podStartE2EDuration="1m28.211862971s" podCreationTimestamp="2025-10-06 08:39:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.211558763 +0000 UTC m=+109.001584343" watchObservedRunningTime="2025-10-06 08:40:58.211862971 +0000 UTC m=+109.001888551" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.212000 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=85.211996415 podStartE2EDuration="1m25.211996415s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.196495395 +0000 UTC m=+108.986520985" watchObservedRunningTime="2025-10-06 08:40:58.211996415 +0000 UTC m=+109.002021995" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.229800 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=61.229777566 podStartE2EDuration="1m1.229777566s" podCreationTimestamp="2025-10-06 08:39:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.229736615 +0000 UTC m=+109.019762285" watchObservedRunningTime="2025-10-06 08:40:58.229777566 +0000 UTC m=+109.019803146" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.244361 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73e8a174-2d08-409d-a302-9d88d17042c0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.244610 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.244736 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.244866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73e8a174-2d08-409d-a302-9d88d17042c0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.245003 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73e8a174-2d08-409d-a302-9d88d17042c0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.286172 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podStartSLOduration=85.286154402 podStartE2EDuration="1m25.286154402s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.285808893 +0000 UTC m=+109.075834503" watchObservedRunningTime="2025-10-06 08:40:58.286154402 +0000 UTC m=+109.076179982" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.329214 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ddvqd" podStartSLOduration=85.329199528 podStartE2EDuration="1m25.329199528s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.328851558 +0000 UTC m=+109.118877158" watchObservedRunningTime="2025-10-06 08:40:58.329199528 +0000 UTC m=+109.119225108" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.345690 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.345749 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.345826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73e8a174-2d08-409d-a302-9d88d17042c0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.345864 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73e8a174-2d08-409d-a302-9d88d17042c0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.345912 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73e8a174-2d08-409d-a302-9d88d17042c0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.346367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.346453 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/73e8a174-2d08-409d-a302-9d88d17042c0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.347434 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/73e8a174-2d08-409d-a302-9d88d17042c0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.349840 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5s7m9" podStartSLOduration=85.349823826 podStartE2EDuration="1m25.349823826s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.348625794 +0000 UTC m=+109.138651424" watchObservedRunningTime="2025-10-06 08:40:58.349823826 +0000 UTC m=+109.139849426" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.354480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73e8a174-2d08-409d-a302-9d88d17042c0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.367291 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/73e8a174-2d08-409d-a302-9d88d17042c0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lkpxd\" (UID: \"73e8a174-2d08-409d-a302-9d88d17042c0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.397752 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=84.397731013 podStartE2EDuration="1m24.397731013s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.397010973 +0000 UTC m=+109.187036573" watchObservedRunningTime="2025-10-06 08:40:58.397731013 +0000 UTC m=+109.187756613" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.423430 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.425817 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7b8zg" podStartSLOduration=85.425790193 podStartE2EDuration="1m25.425790193s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.423601323 +0000 UTC m=+109.213626923" watchObservedRunningTime="2025-10-06 08:40:58.425790193 +0000 UTC m=+109.215815803" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.437318 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bmfk9" podStartSLOduration=84.437298774 podStartE2EDuration="1m24.437298774s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:58.436123792 +0000 UTC m=+109.226149392" watchObservedRunningTime="2025-10-06 08:40:58.437298774 +0000 UTC m=+109.227324364" Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.505538 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" event={"ID":"73e8a174-2d08-409d-a302-9d88d17042c0","Type":"ContainerStarted","Data":"d9fba5f4d8a053558e94194fd130ce4fb7111bb5f94a41f3acd073d4519d6576"} Oct 06 08:40:58 crc kubenswrapper[4989]: I1006 08:40:58.935686 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:40:58 crc kubenswrapper[4989]: E1006 08:40:58.935829 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:40:59 crc kubenswrapper[4989]: I1006 08:40:59.510410 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" event={"ID":"73e8a174-2d08-409d-a302-9d88d17042c0","Type":"ContainerStarted","Data":"ca51b07dad99bd34cc5fa1ab9bd4748719863b9a2f0147d3e5255e6b8a29f5d3"} Oct 06 08:40:59 crc kubenswrapper[4989]: I1006 08:40:59.526179 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lkpxd" podStartSLOduration=86.52615447 podStartE2EDuration="1m26.52615447s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:40:59.525964265 +0000 UTC m=+110.315989885" watchObservedRunningTime="2025-10-06 08:40:59.52615447 +0000 UTC m=+110.316180090" Oct 06 08:40:59 crc kubenswrapper[4989]: I1006 08:40:59.935951 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:40:59 crc kubenswrapper[4989]: I1006 08:40:59.935957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:40:59 crc kubenswrapper[4989]: E1006 08:40:59.936151 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:40:59 crc kubenswrapper[4989]: I1006 08:40:59.936197 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:40:59 crc kubenswrapper[4989]: E1006 08:40:59.936356 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:40:59 crc kubenswrapper[4989]: E1006 08:40:59.936489 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:00 crc kubenswrapper[4989]: I1006 08:41:00.935586 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:00 crc kubenswrapper[4989]: E1006 08:41:00.935751 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:01 crc kubenswrapper[4989]: I1006 08:41:01.935741 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:01 crc kubenswrapper[4989]: I1006 08:41:01.935924 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:01 crc kubenswrapper[4989]: I1006 08:41:01.935951 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:01 crc kubenswrapper[4989]: E1006 08:41:01.936079 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:01 crc kubenswrapper[4989]: E1006 08:41:01.936308 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:01 crc kubenswrapper[4989]: E1006 08:41:01.936426 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:02 crc kubenswrapper[4989]: I1006 08:41:02.935487 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:02 crc kubenswrapper[4989]: E1006 08:41:02.935904 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:02 crc kubenswrapper[4989]: I1006 08:41:02.936080 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:41:02 crc kubenswrapper[4989]: E1006 08:41:02.936753 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:41:03 crc kubenswrapper[4989]: I1006 08:41:03.935253 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:03 crc kubenswrapper[4989]: E1006 08:41:03.935411 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:03 crc kubenswrapper[4989]: I1006 08:41:03.935786 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:03 crc kubenswrapper[4989]: E1006 08:41:03.935902 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:03 crc kubenswrapper[4989]: I1006 08:41:03.936282 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:03 crc kubenswrapper[4989]: E1006 08:41:03.936430 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:04 crc kubenswrapper[4989]: I1006 08:41:04.934971 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:04 crc kubenswrapper[4989]: E1006 08:41:04.935112 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:05 crc kubenswrapper[4989]: I1006 08:41:05.935760 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:05 crc kubenswrapper[4989]: E1006 08:41:05.935959 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:05 crc kubenswrapper[4989]: I1006 08:41:05.936012 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:05 crc kubenswrapper[4989]: I1006 08:41:05.935871 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:05 crc kubenswrapper[4989]: E1006 08:41:05.936221 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:05 crc kubenswrapper[4989]: E1006 08:41:05.936341 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:06 crc kubenswrapper[4989]: I1006 08:41:06.935752 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:06 crc kubenswrapper[4989]: E1006 08:41:06.935869 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:07 crc kubenswrapper[4989]: I1006 08:41:07.935489 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:07 crc kubenswrapper[4989]: I1006 08:41:07.935587 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:07 crc kubenswrapper[4989]: I1006 08:41:07.935486 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:07 crc kubenswrapper[4989]: E1006 08:41:07.935616 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:07 crc kubenswrapper[4989]: E1006 08:41:07.935729 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:07 crc kubenswrapper[4989]: E1006 08:41:07.935908 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:08 crc kubenswrapper[4989]: I1006 08:41:08.935790 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:08 crc kubenswrapper[4989]: E1006 08:41:08.935939 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.541932 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/1.log" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.542589 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/0.log" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.542695 4989 generic.go:334] "Generic (PLEG): container finished" podID="b901f1ca-c342-4faa-a715-0ff3d94bc369" containerID="2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa" exitCode=1 Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.542799 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerDied","Data":"2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa"} Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.542851 4989 scope.go:117] "RemoveContainer" containerID="9535db93c76ce4382e5c8491c926772c192a98f551a97fe50e5c1f4fd8066e69" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.543518 4989 scope.go:117] "RemoveContainer" containerID="2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa" Oct 06 08:41:09 crc kubenswrapper[4989]: E1006 08:41:09.543954 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-wz8dx_openshift-multus(b901f1ca-c342-4faa-a715-0ff3d94bc369)\"" pod="openshift-multus/multus-wz8dx" podUID="b901f1ca-c342-4faa-a715-0ff3d94bc369" Oct 06 08:41:09 crc kubenswrapper[4989]: E1006 08:41:09.883322 4989 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.935826 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.935953 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:09 crc kubenswrapper[4989]: E1006 08:41:09.936800 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:09 crc kubenswrapper[4989]: I1006 08:41:09.936916 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:09 crc kubenswrapper[4989]: E1006 08:41:09.937098 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:09 crc kubenswrapper[4989]: E1006 08:41:09.937192 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:10 crc kubenswrapper[4989]: E1006 08:41:10.044626 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:41:10 crc kubenswrapper[4989]: I1006 08:41:10.550889 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/1.log" Oct 06 08:41:10 crc kubenswrapper[4989]: I1006 08:41:10.935614 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:10 crc kubenswrapper[4989]: E1006 08:41:10.935850 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:11 crc kubenswrapper[4989]: I1006 08:41:11.935614 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:11 crc kubenswrapper[4989]: E1006 08:41:11.935828 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:11 crc kubenswrapper[4989]: I1006 08:41:11.935997 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:11 crc kubenswrapper[4989]: E1006 08:41:11.936240 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:11 crc kubenswrapper[4989]: I1006 08:41:11.936549 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:11 crc kubenswrapper[4989]: E1006 08:41:11.936698 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:12 crc kubenswrapper[4989]: I1006 08:41:12.935829 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:12 crc kubenswrapper[4989]: E1006 08:41:12.936005 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:13 crc kubenswrapper[4989]: I1006 08:41:13.935469 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:13 crc kubenswrapper[4989]: I1006 08:41:13.935513 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:13 crc kubenswrapper[4989]: E1006 08:41:13.936552 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:13 crc kubenswrapper[4989]: E1006 08:41:13.936633 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:13 crc kubenswrapper[4989]: I1006 08:41:13.936744 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:13 crc kubenswrapper[4989]: I1006 08:41:13.936947 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:41:13 crc kubenswrapper[4989]: E1006 08:41:13.937011 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:13 crc kubenswrapper[4989]: E1006 08:41:13.937274 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-t6mdb_openshift-ovn-kubernetes(91d0cca1-dfc6-4f12-a25a-7325782935ba)\"" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" Oct 06 08:41:14 crc kubenswrapper[4989]: I1006 08:41:14.935857 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:14 crc kubenswrapper[4989]: E1006 08:41:14.936031 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:15 crc kubenswrapper[4989]: E1006 08:41:15.046726 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:41:15 crc kubenswrapper[4989]: I1006 08:41:15.935912 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:15 crc kubenswrapper[4989]: I1006 08:41:15.935984 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:15 crc kubenswrapper[4989]: I1006 08:41:15.935935 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:15 crc kubenswrapper[4989]: E1006 08:41:15.936140 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:15 crc kubenswrapper[4989]: E1006 08:41:15.936232 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:15 crc kubenswrapper[4989]: E1006 08:41:15.936330 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:16 crc kubenswrapper[4989]: I1006 08:41:16.936063 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:16 crc kubenswrapper[4989]: E1006 08:41:16.936260 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:17 crc kubenswrapper[4989]: I1006 08:41:17.935275 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:17 crc kubenswrapper[4989]: I1006 08:41:17.935326 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:17 crc kubenswrapper[4989]: I1006 08:41:17.935368 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:17 crc kubenswrapper[4989]: E1006 08:41:17.935508 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:17 crc kubenswrapper[4989]: E1006 08:41:17.935608 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:17 crc kubenswrapper[4989]: E1006 08:41:17.935762 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:18 crc kubenswrapper[4989]: I1006 08:41:18.934878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:18 crc kubenswrapper[4989]: E1006 08:41:18.935296 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:19 crc kubenswrapper[4989]: I1006 08:41:19.935375 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:19 crc kubenswrapper[4989]: I1006 08:41:19.935460 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:19 crc kubenswrapper[4989]: E1006 08:41:19.937170 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:19 crc kubenswrapper[4989]: I1006 08:41:19.937202 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:19 crc kubenswrapper[4989]: E1006 08:41:19.937312 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:19 crc kubenswrapper[4989]: E1006 08:41:19.937583 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:20 crc kubenswrapper[4989]: E1006 08:41:20.047605 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:41:20 crc kubenswrapper[4989]: I1006 08:41:20.935933 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:20 crc kubenswrapper[4989]: E1006 08:41:20.936116 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:21 crc kubenswrapper[4989]: I1006 08:41:21.935793 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:21 crc kubenswrapper[4989]: I1006 08:41:21.935814 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:21 crc kubenswrapper[4989]: I1006 08:41:21.935875 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:21 crc kubenswrapper[4989]: E1006 08:41:21.936015 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:21 crc kubenswrapper[4989]: E1006 08:41:21.936179 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:21 crc kubenswrapper[4989]: E1006 08:41:21.936377 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:22 crc kubenswrapper[4989]: I1006 08:41:22.935693 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:22 crc kubenswrapper[4989]: E1006 08:41:22.935929 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:23 crc kubenswrapper[4989]: I1006 08:41:23.935339 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:23 crc kubenswrapper[4989]: I1006 08:41:23.935437 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:23 crc kubenswrapper[4989]: E1006 08:41:23.935761 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:23 crc kubenswrapper[4989]: I1006 08:41:23.935778 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:23 crc kubenswrapper[4989]: E1006 08:41:23.935874 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:23 crc kubenswrapper[4989]: E1006 08:41:23.935944 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:23 crc kubenswrapper[4989]: I1006 08:41:23.936413 4989 scope.go:117] "RemoveContainer" containerID="2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa" Oct 06 08:41:24 crc kubenswrapper[4989]: I1006 08:41:24.602680 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/1.log" Oct 06 08:41:24 crc kubenswrapper[4989]: I1006 08:41:24.602729 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerStarted","Data":"6cd6e8e428a6d15f8ae746ce43a6eaa2ff4c01b5157090e193f58fe5ac9cdb48"} Oct 06 08:41:24 crc kubenswrapper[4989]: I1006 08:41:24.935930 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:24 crc kubenswrapper[4989]: E1006 08:41:24.936104 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:25 crc kubenswrapper[4989]: E1006 08:41:25.048744 4989 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:41:25 crc kubenswrapper[4989]: I1006 08:41:25.935880 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:25 crc kubenswrapper[4989]: I1006 08:41:25.935929 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:25 crc kubenswrapper[4989]: E1006 08:41:25.936070 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:25 crc kubenswrapper[4989]: I1006 08:41:25.936102 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:25 crc kubenswrapper[4989]: E1006 08:41:25.936219 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:25 crc kubenswrapper[4989]: E1006 08:41:25.937047 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:25 crc kubenswrapper[4989]: I1006 08:41:25.937456 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.610639 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/3.log" Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.613438 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerStarted","Data":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.614347 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.647794 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podStartSLOduration=113.647777753 podStartE2EDuration="1m53.647777753s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:26.647778503 +0000 UTC m=+137.437804083" watchObservedRunningTime="2025-10-06 08:41:26.647777753 +0000 UTC m=+137.437803333" Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.911929 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8ndcb"] Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.912163 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:26 crc kubenswrapper[4989]: E1006 08:41:26.912436 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:26 crc kubenswrapper[4989]: I1006 08:41:26.935920 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:26 crc kubenswrapper[4989]: E1006 08:41:26.936044 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:27 crc kubenswrapper[4989]: I1006 08:41:27.935845 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:27 crc kubenswrapper[4989]: E1006 08:41:27.936034 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:27 crc kubenswrapper[4989]: I1006 08:41:27.936090 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:27 crc kubenswrapper[4989]: E1006 08:41:27.936294 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:28 crc kubenswrapper[4989]: I1006 08:41:28.935465 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:28 crc kubenswrapper[4989]: I1006 08:41:28.935505 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:28 crc kubenswrapper[4989]: E1006 08:41:28.935692 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8ndcb" podUID="c9cca673-1746-4fd9-bc67-330c4d06522e" Oct 06 08:41:28 crc kubenswrapper[4989]: E1006 08:41:28.935918 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 06 08:41:29 crc kubenswrapper[4989]: I1006 08:41:29.935615 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:29 crc kubenswrapper[4989]: I1006 08:41:29.935613 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:29 crc kubenswrapper[4989]: E1006 08:41:29.936906 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 06 08:41:29 crc kubenswrapper[4989]: E1006 08:41:29.937072 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.934961 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.935040 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.938302 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.939540 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.940263 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 06 08:41:30 crc kubenswrapper[4989]: I1006 08:41:30.941213 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 06 08:41:31 crc kubenswrapper[4989]: I1006 08:41:31.935897 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:31 crc kubenswrapper[4989]: I1006 08:41:31.935931 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:31 crc kubenswrapper[4989]: I1006 08:41:31.938518 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 06 08:41:31 crc kubenswrapper[4989]: I1006 08:41:31.939163 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.784097 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:37 crc kubenswrapper[4989]: E1006 08:41:37.784255 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:43:39.784226176 +0000 UTC m=+270.574251766 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.784302 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.784376 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.784446 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.784467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.785962 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.790010 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.790366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.791755 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.871702 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 06 08:41:37 crc kubenswrapper[4989]: I1006 08:41:37.956136 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.007917 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:38 crc kubenswrapper[4989]: W1006 08:41:38.130923 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-36dce5f1c02601c0e168600cc374bf8cbc7dcfa7ea8f519394c0eb6dda1f0593 WatchSource:0}: Error finding container 36dce5f1c02601c0e168600cc374bf8cbc7dcfa7ea8f519394c0eb6dda1f0593: Status 404 returned error can't find the container with id 36dce5f1c02601c0e168600cc374bf8cbc7dcfa7ea8f519394c0eb6dda1f0593 Oct 06 08:41:38 crc kubenswrapper[4989]: W1006 08:41:38.170782 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-865fc8e4c4f64de8f74f9b68748c94461f63acb2cfb94c001d508f10c88991b7 WatchSource:0}: Error finding container 865fc8e4c4f64de8f74f9b68748c94461f63acb2cfb94c001d508f10c88991b7: Status 404 returned error can't find the container with id 865fc8e4c4f64de8f74f9b68748c94461f63acb2cfb94c001d508f10c88991b7 Oct 06 08:41:38 crc kubenswrapper[4989]: W1006 08:41:38.197015 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-ace5ff6958ec974f6a0da78afa5625a385137716e4c3b1d91e39726caa973004 WatchSource:0}: Error finding container ace5ff6958ec974f6a0da78afa5625a385137716e4c3b1d91e39726caa973004: Status 404 returned error can't find the container with id ace5ff6958ec974f6a0da78afa5625a385137716e4c3b1d91e39726caa973004 Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.661389 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0ae09f478c28550fa03e960fe765e594f0cb06fbbbab749a2a5aec1ecde8bc46"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.661463 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ace5ff6958ec974f6a0da78afa5625a385137716e4c3b1d91e39726caa973004"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.661739 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.663824 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2497136bf492fbc843d902697212f4302e31b0ae388a903250a01b3d266c39b1"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.664094 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"865fc8e4c4f64de8f74f9b68748c94461f63acb2cfb94c001d508f10c88991b7"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.667145 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6463bad20a876b5c9d5ff3957f480f46b0832ef59d912fc2c29fcdd964bc60c9"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.667226 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"36dce5f1c02601c0e168600cc374bf8cbc7dcfa7ea8f519394c0eb6dda1f0593"} Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.830626 4989 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.864722 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-87qfj"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.865372 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.870303 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.870937 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.871018 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.871321 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.871834 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.871974 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.872729 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.873578 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.873899 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.874253 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.874292 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.874364 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.888499 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.892574 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.893030 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.893315 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.893530 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.893624 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.897335 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.897968 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.898812 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.898964 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899127 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899280 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899379 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899730 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899995 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900047 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900190 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900325 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900465 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900622 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.900986 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.901202 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.901523 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.899999 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.902048 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.906796 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9pfhc"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.907426 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.908125 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.929390 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.929615 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.929823 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.932327 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.933013 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.936136 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.936334 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.936500 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.937197 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.937398 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.937908 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.943698 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.944010 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.944207 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.944251 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.944377 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.945317 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.948563 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.959501 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.960445 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wm56d"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.961388 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.962202 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.963003 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.963271 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.964519 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.965364 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rv274"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.966455 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.966753 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97"] Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.983578 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 06 08:41:38 crc kubenswrapper[4989]: I1006 08:41:38.984142 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.007006 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.007152 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.008962 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b2g98"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.009593 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.009928 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcq7v"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.010403 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.010846 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.010866 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.011146 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.011457 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.011493 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.011646 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.011733 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.015760 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.016180 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.016841 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.017545 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.017612 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.018478 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.018860 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.019308 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.019425 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.019957 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69dw"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.020712 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021322 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/118c9ba6-fc48-4427-bf18-82c4ec482e20-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021357 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4tpc\" (UniqueName: \"kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021383 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-audit-dir\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021401 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f085dab8-7067-4be2-844c-884ba9688549-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021418 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021436 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021458 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-policies\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021475 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-client\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021496 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021516 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-dir\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sph6m\" (UniqueName: \"kubernetes.io/projected/f085dab8-7067-4be2-844c-884ba9688549-kube-api-access-sph6m\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021558 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021581 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021604 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021623 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-machine-approver-tls\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021642 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021681 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021700 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021721 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx5kh\" (UniqueName: \"kubernetes.io/projected/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-kube-api-access-dx5kh\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021743 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021763 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021781 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-serving-cert\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021803 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-serving-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021822 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69bcj\" (UniqueName: \"kubernetes.io/projected/61635015-0a25-4834-8f24-2ecee7ee9240-kube-api-access-69bcj\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021855 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f085dab8-7067-4be2-844c-884ba9688549-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021876 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021897 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-auth-proxy-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021918 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/595af3bc-3f4a-4214-a46b-4ea6b57136e2-serving-cert\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021961 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-config\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.021980 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022000 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022019 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-encryption-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022039 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022058 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022078 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-serving-cert\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022098 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmgjk\" (UniqueName: \"kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-676s6\" (UniqueName: \"kubernetes.io/projected/3bd79c86-9633-4dcb-bc5f-800abe66a513-kube-api-access-676s6\") pod \"downloads-7954f5f757-9pfhc\" (UID: \"3bd79c86-9633-4dcb-bc5f-800abe66a513\") " pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022140 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-image-import-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022182 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95qfr\" (UniqueName: \"kubernetes.io/projected/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-kube-api-access-95qfr\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022202 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-trusted-ca-bundle\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022225 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022244 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-trusted-ca\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022283 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn46v\" (UniqueName: \"kubernetes.io/projected/595af3bc-3f4a-4214-a46b-4ea6b57136e2-kube-api-access-kn46v\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022307 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sj2n\" (UniqueName: \"kubernetes.io/projected/118c9ba6-fc48-4427-bf18-82c4ec482e20-kube-api-access-7sj2n\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022327 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-audit\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022347 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022367 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-node-pullsecrets\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022389 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022412 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-client\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022438 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.022461 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-encryption-config\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029033 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029149 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029114 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029463 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029706 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.029989 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-m7lpj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.030632 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.031055 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.031103 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.031186 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.033242 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.033648 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.033741 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.033958 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.034019 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.034731 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.035214 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.035685 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.061724 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.062126 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.062511 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.063074 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.063494 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.064161 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlmtd"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.064411 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.065600 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.068864 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.069731 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.069813 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.069864 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.069934 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.070168 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.080576 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.080666 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.080684 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.080914 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.085733 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086211 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086279 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086352 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086550 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086671 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086707 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086805 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.086817 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090187 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090368 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090615 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090663 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090781 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-87qfj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090890 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090913 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090956 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091033 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091129 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.090894 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091264 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091355 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091491 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.091797 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.092197 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.093460 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.094072 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.094747 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.097760 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.098195 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.098858 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.098887 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.099311 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.099723 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.099828 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.100006 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.100187 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.101692 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8zkhj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.102030 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.102270 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.102350 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.103149 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.106115 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.107843 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.108370 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.108724 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.110267 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.116238 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.116691 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.117170 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.117253 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.117415 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.117706 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.118706 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-c829n"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.119300 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.119728 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.120328 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.120931 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.122038 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9pfhc"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123104 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123359 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/118c9ba6-fc48-4427-bf18-82c4ec482e20-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4tpc\" (UniqueName: \"kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-audit-dir\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f085dab8-7067-4be2-844c-884ba9688549-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123432 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123449 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-policies\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123482 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-client\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123497 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-dir\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123524 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sph6m\" (UniqueName: \"kubernetes.io/projected/f085dab8-7067-4be2-844c-884ba9688549-kube-api-access-sph6m\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123537 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123553 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123590 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123604 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-machine-approver-tls\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123708 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97d62f70-c96b-4468-bc7f-c110115a03b6-tmpfs\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ns7v\" (UniqueName: \"kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123746 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123761 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123777 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx5kh\" (UniqueName: \"kubernetes.io/projected/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-kube-api-access-dx5kh\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123807 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123834 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123849 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-serving-cert\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123864 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-serving-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123889 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69bcj\" (UniqueName: \"kubernetes.io/projected/61635015-0a25-4834-8f24-2ecee7ee9240-kube-api-access-69bcj\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f085dab8-7067-4be2-844c-884ba9688549-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123946 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123964 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-auth-proxy-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123968 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4wcbb"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123977 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/595af3bc-3f4a-4214-a46b-4ea6b57136e2-serving-cert\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.123991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124104 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-config\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-encryption-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124243 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-webhook-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124273 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgbnj\" (UniqueName: \"kubernetes.io/projected/97d62f70-c96b-4468-bc7f-c110115a03b6-kube-api-access-bgbnj\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124296 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124347 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-serving-cert\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124405 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmgjk\" (UniqueName: \"kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124435 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-676s6\" (UniqueName: \"kubernetes.io/projected/3bd79c86-9633-4dcb-bc5f-800abe66a513-kube-api-access-676s6\") pod \"downloads-7954f5f757-9pfhc\" (UID: \"3bd79c86-9633-4dcb-bc5f-800abe66a513\") " pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124498 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124530 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124586 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-image-import-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124614 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95qfr\" (UniqueName: \"kubernetes.io/projected/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-kube-api-access-95qfr\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124642 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-trusted-ca-bundle\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124821 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124854 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-trusted-ca\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124874 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn46v\" (UniqueName: \"kubernetes.io/projected/595af3bc-3f4a-4214-a46b-4ea6b57136e2-kube-api-access-kn46v\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124893 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sj2n\" (UniqueName: \"kubernetes.io/projected/118c9ba6-fc48-4427-bf18-82c4ec482e20-kube-api-access-7sj2n\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-audit\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124967 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124981 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-node-pullsecrets\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.124995 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.125012 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.125020 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126507 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlmtd"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126813 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126847 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-client\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126881 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126915 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-encryption-config\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.126973 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.128364 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-audit-dir\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.128923 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-client\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.129444 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-policies\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.131912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.132005 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.132461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.132993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.141496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/61635015-0a25-4834-8f24-2ecee7ee9240-node-pullsecrets\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.142511 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.142819 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-audit\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.142994 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.143005 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.143063 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.143437 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.143811 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.145319 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.145789 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.145802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.145899 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-m7lpj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.145952 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.146429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.148122 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.149885 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.153518 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-encryption-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.153797 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f085dab8-7067-4be2-844c-884ba9688549-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.154361 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/118c9ba6-fc48-4427-bf18-82c4ec482e20-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.154665 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-serving-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.155006 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f085dab8-7067-4be2-844c-884ba9688549-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.155099 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.155194 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-image-import-ca\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.155612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-config\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.155647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61635015-0a25-4834-8f24-2ecee7ee9240-trusted-ca-bundle\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.156108 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.156982 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.157010 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.157164 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-audit-dir\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.157627 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-auth-proxy-config\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.157776 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-serving-cert\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158139 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158281 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-serving-cert\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158698 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61635015-0a25-4834-8f24-2ecee7ee9240-etcd-client\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158753 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-config\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158801 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/595af3bc-3f4a-4214-a46b-4ea6b57136e2-serving-cert\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158923 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/595af3bc-3f4a-4214-a46b-4ea6b57136e2-trusted-ca\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.158936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.159017 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.161254 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-encryption-config\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.163500 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-machine-approver-tls\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.164963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.167169 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.170675 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.172214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.173512 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.176030 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6qqlb"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.176847 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.178486 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69dw"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.180200 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.181506 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.182897 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.184353 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.185759 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.187521 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.188128 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.189235 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcq7v"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.190449 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.191592 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.193219 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.194267 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wm56d"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.195366 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rv274"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.196546 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.197880 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8zkhj"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.199014 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b2g98"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.200330 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.201468 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.202861 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4wcbb"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.203858 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.204892 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.206924 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.207206 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.208283 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x9rxz"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.209361 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zlrl6"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.209502 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.210730 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x9rxz"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.210859 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.211475 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zlrl6"] Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227401 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-webhook-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgbnj\" (UniqueName: \"kubernetes.io/projected/97d62f70-c96b-4468-bc7f-c110115a03b6-kube-api-access-bgbnj\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227555 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227690 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227773 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97d62f70-c96b-4468-bc7f-c110115a03b6-tmpfs\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.227801 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ns7v\" (UniqueName: \"kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.228040 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.228571 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/97d62f70-c96b-4468-bc7f-c110115a03b6-tmpfs\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.229211 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.229446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.230932 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.247839 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.273372 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.287592 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.308608 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.329515 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.348574 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.368121 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.388202 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.407242 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.428165 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.447503 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.467806 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.489255 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.508612 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.529975 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.547626 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.568198 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.587638 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.592750 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-webhook-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.593189 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/97d62f70-c96b-4468-bc7f-c110115a03b6-apiservice-cert\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.608482 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.629369 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.648982 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.668938 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.688106 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.708200 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.728602 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.747932 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.767553 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.788496 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.808044 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.828114 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.851091 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.867295 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.887810 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.908714 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.928437 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.947741 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.967406 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 06 08:41:39 crc kubenswrapper[4989]: I1006 08:41:39.987438 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.014133 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.027765 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.048248 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.067922 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.088021 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.106950 4989 request.go:700] Waited for 1.004353249s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/secrets?fieldSelector=metadata.name%3Detcd-operator-dockercfg-r9srn&limit=500&resourceVersion=0 Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.108358 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.128340 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.147694 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.168496 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.188634 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.208113 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.229000 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.248336 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.268509 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.288302 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.308929 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.328990 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.348358 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.369242 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.388885 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.408061 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.428931 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.448893 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.468234 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.488517 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.508582 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.528534 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.548289 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.568757 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.588674 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.608465 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.628872 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.647777 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.668108 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.688261 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.709445 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.728999 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.748248 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.798931 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4tpc\" (UniqueName: \"kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc\") pod \"oauth-openshift-558db77b4-krvc2\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.812166 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmgjk\" (UniqueName: \"kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk\") pod \"console-f9d7485db-wwr9w\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.836644 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx5kh\" (UniqueName: \"kubernetes.io/projected/ee383275-f8d1-4dca-89e5-978bfa5ad4e7-kube-api-access-dx5kh\") pod \"apiserver-7bbb656c7d-rrqq6\" (UID: \"ee383275-f8d1-4dca-89e5-978bfa5ad4e7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.844839 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-676s6\" (UniqueName: \"kubernetes.io/projected/3bd79c86-9633-4dcb-bc5f-800abe66a513-kube-api-access-676s6\") pod \"downloads-7954f5f757-9pfhc\" (UID: \"3bd79c86-9633-4dcb-bc5f-800abe66a513\") " pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.860764 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95qfr\" (UniqueName: \"kubernetes.io/projected/057c0c57-faf9-439b-b8a1-f5f4a659f0d8-kube-api-access-95qfr\") pod \"machine-approver-56656f9798-mtfp9\" (UID: \"057c0c57-faf9-439b-b8a1-f5f4a659f0d8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.879871 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sj2n\" (UniqueName: \"kubernetes.io/projected/118c9ba6-fc48-4427-bf18-82c4ec482e20-kube-api-access-7sj2n\") pod \"cluster-samples-operator-665b6dd947-6t7bd\" (UID: \"118c9ba6-fc48-4427-bf18-82c4ec482e20\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.905561 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn46v\" (UniqueName: \"kubernetes.io/projected/595af3bc-3f4a-4214-a46b-4ea6b57136e2-kube-api-access-kn46v\") pod \"console-operator-58897d9998-wm56d\" (UID: \"595af3bc-3f4a-4214-a46b-4ea6b57136e2\") " pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.919529 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69bcj\" (UniqueName: \"kubernetes.io/projected/61635015-0a25-4834-8f24-2ecee7ee9240-kube-api-access-69bcj\") pod \"apiserver-76f77b778f-87qfj\" (UID: \"61635015-0a25-4834-8f24-2ecee7ee9240\") " pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.939487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sph6m\" (UniqueName: \"kubernetes.io/projected/f085dab8-7067-4be2-844c-884ba9688549-kube-api-access-sph6m\") pod \"openshift-apiserver-operator-796bbdcf4f-r96zs\" (UID: \"f085dab8-7067-4be2-844c-884ba9688549\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954523 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954586 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvmxk\" (UniqueName: \"kubernetes.io/projected/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-kube-api-access-vvmxk\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954631 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-config\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954695 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rrhl\" (UniqueName: \"kubernetes.io/projected/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-kube-api-access-6rrhl\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954720 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954771 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk4lt\" (UniqueName: \"kubernetes.io/projected/92c1ecab-13ee-4b81-9454-ad02e7071592-kube-api-access-bk4lt\") pod \"migrator-59844c95c7-fx7cq\" (UID: \"92c1ecab-13ee-4b81-9454-ad02e7071592\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954792 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954839 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/281b11d6-f406-4609-a432-5a796b2304ca-signing-cabundle\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954893 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954921 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954948 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m88l\" (UniqueName: \"kubernetes.io/projected/9ec320c3-776d-4197-9259-7feee562229a-kube-api-access-5m88l\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.954982 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955002 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjb7s\" (UniqueName: \"kubernetes.io/projected/51938ce0-b08d-456a-b711-b3de25465a9d-kube-api-access-kjb7s\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955043 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-config\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955199 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ec320c3-776d-4197-9259-7feee562229a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955217 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51938ce0-b08d-456a-b711-b3de25465a9d-metrics-tls\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955231 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcshv\" (UniqueName: \"kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955253 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c52m\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-kube-api-access-9c52m\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955278 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59bnt\" (UniqueName: \"kubernetes.io/projected/fc8b7232-28ce-4453-a5cd-eddb6077f07d-kube-api-access-59bnt\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955291 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa0ddc12-b6aa-4159-9945-042260a0595c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955328 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: E1006 08:41:40.955361 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.455344882 +0000 UTC m=+152.245370462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955399 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955441 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7dcad5af-901c-431f-b7b9-a98b4e667ee8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955460 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955495 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtzp8\" (UniqueName: \"kubernetes.io/projected/72aca57c-867e-460f-b3c9-8d0a2bc275ca-kube-api-access-gtzp8\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955524 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955543 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-images\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955569 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955587 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955600 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/281b11d6-f406-4609-a432-5a796b2304ca-signing-key\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955615 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955629 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt5t2\" (UniqueName: \"kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955646 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955747 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b7232-28ce-4453-a5cd-eddb6077f07d-serving-cert\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955795 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-srv-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7hck\" (UniqueName: \"kubernetes.io/projected/7dcad5af-901c-431f-b7b9-a98b4e667ee8-kube-api-access-m7hck\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955828 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955857 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbz42\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955885 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsprv\" (UniqueName: \"kubernetes.io/projected/6d858cfb-196c-40b5-8b67-de2529a4d3c4-kube-api-access-fsprv\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955940 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.955981 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-srv-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-serving-cert\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956086 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa0ddc12-b6aa-4159-9945-042260a0595c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh4q4\" (UniqueName: \"kubernetes.io/projected/281b11d6-f406-4609-a432-5a796b2304ca-kube-api-access-kh4q4\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956144 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72aca57c-867e-460f-b3c9-8d0a2bc275ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956171 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwdjb\" (UniqueName: \"kubernetes.io/projected/8bce9738-a274-4c30-913e-37ea8663bee6-kube-api-access-nwdjb\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956194 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.956215 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8bce9738-a274-4c30-913e-37ea8663bee6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.984525 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.988169 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 06 08:41:40 crc kubenswrapper[4989]: I1006 08:41:40.996375 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.005051 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.007465 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.017885 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.027273 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.028855 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.035957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.048449 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056708 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056888 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-node-bootstrap-token\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056922 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcshv\" (UniqueName: \"kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056945 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b702f6-d926-4d3e-b509-60facd64ed7c-config\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056975 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-proxy-tls\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.056998 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a46506a-7ebe-405b-8ff1-88a9122390e6-metrics-tls\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057016 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-proxy-tls\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057036 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c52m\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-kube-api-access-9c52m\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057061 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59bnt\" (UniqueName: \"kubernetes.io/projected/fc8b7232-28ce-4453-a5cd-eddb6077f07d-kube-api-access-59bnt\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057081 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057103 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b2a364f-6a58-442e-9dc8-3807d92943e3-config\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87j2g\" (UniqueName: \"kubernetes.io/projected/0a46506a-7ebe-405b-8ff1-88a9122390e6-kube-api-access-87j2g\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057143 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057166 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7dcad5af-901c-431f-b7b9-a98b4e667ee8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057221 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057242 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a46506a-7ebe-405b-8ff1-88a9122390e6-config-volume\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057286 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h8dw\" (UniqueName: \"kubernetes.io/projected/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-kube-api-access-5h8dw\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057311 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-images\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057356 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/868aae56-7e7a-4af7-8c31-b97448d5c81f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057375 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a05a51e-c77e-4272-91fa-290cbec0f3d3-cert\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057396 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt5t2\" (UniqueName: \"kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-images\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057446 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87042d1-b2c4-44ac-9f0c-853b88375f53-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057465 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-mountpoint-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057484 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-serving-cert\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057503 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-srv-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057539 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa0ddc12-b6aa-4159-9945-042260a0595c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057578 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvmxk\" (UniqueName: \"kubernetes.io/projected/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-kube-api-access-vvmxk\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057603 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dkdk\" (UniqueName: \"kubernetes.io/projected/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-kube-api-access-2dkdk\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057628 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr648\" (UniqueName: \"kubernetes.io/projected/4a05a51e-c77e-4272-91fa-290cbec0f3d3-kube-api-access-lr648\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057701 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b2a364f-6a58-442e-9dc8-3807d92943e3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057725 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-metrics-certs\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057752 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f81572-1755-48a1-b822-b13798b463a4-trusted-ca\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057806 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flbj2\" (UniqueName: \"kubernetes.io/projected/a52e7f91-d84a-495d-a6ef-dcd0d0618990-kube-api-access-flbj2\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057830 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk4lt\" (UniqueName: \"kubernetes.io/projected/92c1ecab-13ee-4b81-9454-ad02e7071592-kube-api-access-bk4lt\") pod \"migrator-59844c95c7-fx7cq\" (UID: \"92c1ecab-13ee-4b81-9454-ad02e7071592\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057852 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/281b11d6-f406-4609-a432-5a796b2304ca-signing-cabundle\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057875 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057901 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd7241b3-d77b-458e-b4bf-77fbd9f28596-service-ca-bundle\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057927 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m88l\" (UniqueName: \"kubernetes.io/projected/9ec320c3-776d-4197-9259-7feee562229a-kube-api-access-5m88l\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057948 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057971 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-socket-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.057994 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-registration-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058019 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ec320c3-776d-4197-9259-7feee562229a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058042 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51938ce0-b08d-456a-b711-b3de25465a9d-metrics-tls\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058067 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-config\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058090 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xgmg\" (UniqueName: \"kubernetes.io/projected/bd7241b3-d77b-458e-b4bf-77fbd9f28596-kube-api-access-6xgmg\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-csi-data-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058137 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gppxq\" (UniqueName: \"kubernetes.io/projected/c76e2108-d346-4eeb-93b7-1180c2617371-kube-api-access-gppxq\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058158 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c7313b-b022-44fc-bb7f-b35c3fea82de-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058191 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa0ddc12-b6aa-4159-9945-042260a0595c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058215 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4b702f6-d926-4d3e-b509-60facd64ed7c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmkc9\" (UniqueName: \"kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-client\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058284 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b2a364f-6a58-442e-9dc8-3807d92943e3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtzp8\" (UniqueName: \"kubernetes.io/projected/72aca57c-867e-460f-b3c9-8d0a2bc275ca-kube-api-access-gtzp8\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058363 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87042d1-b2c4-44ac-9f0c-853b88375f53-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058391 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-config\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058438 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7313b-b022-44fc-bb7f-b35c3fea82de-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058458 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c7313b-b022-44fc-bb7f-b35c3fea82de-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058482 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/281b11d6-f406-4609-a432-5a796b2304ca-signing-key\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058505 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058555 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058592 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4b702f6-d926-4d3e-b509-60facd64ed7c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl2qg\" (UniqueName: \"kubernetes.io/projected/a87042d1-b2c4-44ac-9f0c-853b88375f53-kube-api-access-rl2qg\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058639 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058682 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b7232-28ce-4453-a5cd-eddb6077f07d-serving-cert\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058707 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-default-certificate\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058733 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7hck\" (UniqueName: \"kubernetes.io/projected/7dcad5af-901c-431f-b7b9-a98b4e667ee8-kube-api-access-m7hck\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058757 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058779 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058803 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbz42\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058827 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsprv\" (UniqueName: \"kubernetes.io/projected/6d858cfb-196c-40b5-8b67-de2529a4d3c4-kube-api-access-fsprv\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058850 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41f81572-1755-48a1-b822-b13798b463a4-metrics-tls\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058872 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktcrq\" (UniqueName: \"kubernetes.io/projected/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-kube-api-access-ktcrq\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058897 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt7tp\" (UniqueName: \"kubernetes.io/projected/868aae56-7e7a-4af7-8c31-b97448d5c81f-kube-api-access-qt7tp\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-srv-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058944 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvsjx\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-kube-api-access-zvsjx\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058969 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-serving-cert\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.058996 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh4q4\" (UniqueName: \"kubernetes.io/projected/281b11d6-f406-4609-a432-5a796b2304ca-kube-api-access-kh4q4\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059021 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72aca57c-867e-460f-b3c9-8d0a2bc275ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059046 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwdjb\" (UniqueName: \"kubernetes.io/projected/8bce9738-a274-4c30-913e-37ea8663bee6-kube-api-access-nwdjb\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059070 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8bce9738-a274-4c30-913e-37ea8663bee6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/868aae56-7e7a-4af7-8c31-b97448d5c81f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059140 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-serving-cert\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059165 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059189 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-config\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059200 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059211 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rrhl\" (UniqueName: \"kubernetes.io/projected/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-kube-api-access-6rrhl\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059281 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-service-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059328 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059348 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-certs\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059367 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-kube-api-access-4kf8m\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059411 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059454 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-stats-auth\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059519 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-plugins-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059584 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjb7s\" (UniqueName: \"kubernetes.io/projected/51938ce0-b08d-456a-b711-b3de25465a9d-kube-api-access-kjb7s\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059608 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.059664 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-config\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.060473 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-config\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.060550 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.560536089 +0000 UTC m=+152.350561669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.061281 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.061874 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.061965 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.061978 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.063554 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.063585 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/7dcad5af-901c-431f-b7b9-a98b4e667ee8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.064055 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/281b11d6-f406-4609-a432-5a796b2304ca-signing-cabundle\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.064347 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.064511 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc8b7232-28ce-4453-a5cd-eddb6077f07d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.064592 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc8b7232-28ce-4453-a5cd-eddb6077f07d-serving-cert\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.065219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.065690 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-profile-collector-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.066497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aa0ddc12-b6aa-4159-9945-042260a0595c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.067142 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-config\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.068177 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa0ddc12-b6aa-4159-9945-042260a0595c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.068700 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9ec320c3-776d-4197-9259-7feee562229a-images\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.069212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.069244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.069580 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.070336 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.070518 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-srv-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.070555 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-srv-cert\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.070574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-serving-cert\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.071513 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ec320c3-776d-4197-9259-7feee562229a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.072205 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/72aca57c-867e-460f-b3c9-8d0a2bc275ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.072348 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8bce9738-a274-4c30-913e-37ea8663bee6-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.072536 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.072875 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/281b11d6-f406-4609-a432-5a796b2304ca-signing-key\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.074972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.075291 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d858cfb-196c-40b5-8b67-de2529a4d3c4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.075794 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.076254 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.083076 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51938ce0-b08d-456a-b711-b3de25465a9d-metrics-tls\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.089229 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.108897 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.112954 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.129576 4989 request.go:700] Waited for 1.918494365s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.131572 4989 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.148817 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.160965 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-serving-cert\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dkdk\" (UniqueName: \"kubernetes.io/projected/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-kube-api-access-2dkdk\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161051 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr648\" (UniqueName: \"kubernetes.io/projected/4a05a51e-c77e-4272-91fa-290cbec0f3d3-kube-api-access-lr648\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161074 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b2a364f-6a58-442e-9dc8-3807d92943e3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-metrics-certs\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161119 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f81572-1755-48a1-b822-b13798b463a4-trusted-ca\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flbj2\" (UniqueName: \"kubernetes.io/projected/a52e7f91-d84a-495d-a6ef-dcd0d0618990-kube-api-access-flbj2\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161171 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd7241b3-d77b-458e-b4bf-77fbd9f28596-service-ca-bundle\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161199 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161219 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-socket-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161242 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-registration-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161262 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-config\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161283 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xgmg\" (UniqueName: \"kubernetes.io/projected/bd7241b3-d77b-458e-b4bf-77fbd9f28596-kube-api-access-6xgmg\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161306 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-csi-data-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gppxq\" (UniqueName: \"kubernetes.io/projected/c76e2108-d346-4eeb-93b7-1180c2617371-kube-api-access-gppxq\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161356 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c7313b-b022-44fc-bb7f-b35c3fea82de-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161389 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4b702f6-d926-4d3e-b509-60facd64ed7c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161411 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmkc9\" (UniqueName: \"kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-client\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161455 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b2a364f-6a58-442e-9dc8-3807d92943e3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161487 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87042d1-b2c4-44ac-9f0c-853b88375f53-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161516 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-config\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161540 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161563 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c7313b-b022-44fc-bb7f-b35c3fea82de-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161586 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7313b-b022-44fc-bb7f-b35c3fea82de-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161613 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4b702f6-d926-4d3e-b509-60facd64ed7c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161632 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl2qg\" (UniqueName: \"kubernetes.io/projected/a87042d1-b2c4-44ac-9f0c-853b88375f53-kube-api-access-rl2qg\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161674 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-default-certificate\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161705 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41f81572-1755-48a1-b822-b13798b463a4-metrics-tls\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161753 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktcrq\" (UniqueName: \"kubernetes.io/projected/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-kube-api-access-ktcrq\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161786 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt7tp\" (UniqueName: \"kubernetes.io/projected/868aae56-7e7a-4af7-8c31-b97448d5c81f-kube-api-access-qt7tp\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161809 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvsjx\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-kube-api-access-zvsjx\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161856 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/868aae56-7e7a-4af7-8c31-b97448d5c81f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-serving-cert\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161908 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-service-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161936 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-certs\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.161981 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-kube-api-access-4kf8m\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162006 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-stats-auth\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162028 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-plugins-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162056 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162088 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162110 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-node-bootstrap-token\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162144 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b702f6-d926-4d3e-b509-60facd64ed7c-config\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162164 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-proxy-tls\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162186 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-proxy-tls\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a46506a-7ebe-405b-8ff1-88a9122390e6-metrics-tls\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162241 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b2a364f-6a58-442e-9dc8-3807d92943e3-config\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162289 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87j2g\" (UniqueName: \"kubernetes.io/projected/0a46506a-7ebe-405b-8ff1-88a9122390e6-kube-api-access-87j2g\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162314 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162338 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a46506a-7ebe-405b-8ff1-88a9122390e6-config-volume\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162358 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h8dw\" (UniqueName: \"kubernetes.io/projected/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-kube-api-access-5h8dw\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162388 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/868aae56-7e7a-4af7-8c31-b97448d5c81f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a05a51e-c77e-4272-91fa-290cbec0f3d3-cert\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162439 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-images\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87042d1-b2c4-44ac-9f0c-853b88375f53-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-mountpoint-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.162595 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-mountpoint-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.163702 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.168139 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/41f81572-1755-48a1-b822-b13798b463a4-metrics-tls\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.168217 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-service-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.168759 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.168898 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/868aae56-7e7a-4af7-8c31-b97448d5c81f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.169351 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-plugins-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.169375 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-serving-cert\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.169693 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.669646104 +0000 UTC m=+152.459671754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.170317 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.171563 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41f81572-1755-48a1-b822-b13798b463a4-trusted-ca\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.173595 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-node-bootstrap-token\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.174292 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4b702f6-d926-4d3e-b509-60facd64ed7c-config\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.174698 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-client\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.174988 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd7241b3-d77b-458e-b4bf-77fbd9f28596-service-ca-bundle\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.175355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-config\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.175413 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87042d1-b2c4-44ac-9f0c-853b88375f53-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.175448 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-stats-auth\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.176555 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b2a364f-6a58-442e-9dc8-3807d92943e3-config\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.176580 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a46506a-7ebe-405b-8ff1-88a9122390e6-config-volume\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.176834 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-socket-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.176901 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-registration-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.177265 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-etcd-ca\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.177513 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-config\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.177563 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14c7313b-b022-44fc-bb7f-b35c3fea82de-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.177805 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a52e7f91-d84a-495d-a6ef-dcd0d0618990-csi-data-dir\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.179481 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-serving-cert\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.179965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87042d1-b2c4-44ac-9f0c-853b88375f53-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.180113 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-metrics-certs\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.180272 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14c7313b-b022-44fc-bb7f-b35c3fea82de-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.180395 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-proxy-tls\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.180764 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-images\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.180970 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-proxy-tls\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.181123 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4a05a51e-c77e-4272-91fa-290cbec0f3d3-cert\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.182283 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd7241b3-d77b-458e-b4bf-77fbd9f28596-default-certificate\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.182629 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c76e2108-d346-4eeb-93b7-1180c2617371-certs\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.182947 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4b702f6-d926-4d3e-b509-60facd64ed7c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.183405 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0a46506a-7ebe-405b-8ff1-88a9122390e6-metrics-tls\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.183999 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/868aae56-7e7a-4af7-8c31-b97448d5c81f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.184299 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b2a364f-6a58-442e-9dc8-3807d92943e3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.185410 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.193111 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ns7v\" (UniqueName: \"kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v\") pod \"route-controller-manager-6576b87f9c-ww9xn\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.219516 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgbnj\" (UniqueName: \"kubernetes.io/projected/97d62f70-c96b-4468-bc7f-c110115a03b6-kube-api-access-bgbnj\") pod \"packageserver-d55dfcdfc-2cwvl\" (UID: \"97d62f70-c96b-4468-bc7f-c110115a03b6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.226000 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.240772 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rrhl\" (UniqueName: \"kubernetes.io/projected/bf7a30ae-66ce-49b6-b59b-cdf42a8662ff-kube-api-access-6rrhl\") pod \"catalog-operator-68c6474976-ns27q\" (UID: \"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.263833 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.264428 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.764408878 +0000 UTC m=+152.554434458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.281005 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcshv\" (UniqueName: \"kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv\") pod \"controller-manager-879f6c89f-vhcsz\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.283129 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c52m\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-kube-api-access-9c52m\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.304955 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59bnt\" (UniqueName: \"kubernetes.io/projected/fc8b7232-28ce-4453-a5cd-eddb6077f07d-kube-api-access-59bnt\") pod \"authentication-operator-69f744f599-fcq7v\" (UID: \"fc8b7232-28ce-4453-a5cd-eddb6077f07d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.326341 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.330492 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m88l\" (UniqueName: \"kubernetes.io/projected/9ec320c3-776d-4197-9259-7feee562229a-kube-api-access-5m88l\") pod \"machine-api-operator-5694c8668f-m7lpj\" (UID: \"9ec320c3-776d-4197-9259-7feee562229a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.344939 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.348143 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk4lt\" (UniqueName: \"kubernetes.io/projected/92c1ecab-13ee-4b81-9454-ad02e7071592-kube-api-access-bk4lt\") pod \"migrator-59844c95c7-fx7cq\" (UID: \"92c1ecab-13ee-4b81-9454-ad02e7071592\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.350227 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.350981 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9pfhc"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.365165 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsprv\" (UniqueName: \"kubernetes.io/projected/6d858cfb-196c-40b5-8b67-de2529a4d3c4-kube-api-access-fsprv\") pod \"olm-operator-6b444d44fb-s2rlf\" (UID: \"6d858cfb-196c-40b5-8b67-de2529a4d3c4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.365708 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.366002 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.865990069 +0000 UTC m=+152.656015649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.382423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa0ddc12-b6aa-4159-9945-042260a0595c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xql97\" (UID: \"aa0ddc12-b6aa-4159-9945-042260a0595c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.405483 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtzp8\" (UniqueName: \"kubernetes.io/projected/72aca57c-867e-460f-b3c9-8d0a2bc275ca-kube-api-access-gtzp8\") pod \"package-server-manager-789f6589d5-2h8tc\" (UID: \"72aca57c-867e-460f-b3c9-8d0a2bc275ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.411991 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.421318 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7hck\" (UniqueName: \"kubernetes.io/projected/7dcad5af-901c-431f-b7b9-a98b4e667ee8-kube-api-access-m7hck\") pod \"control-plane-machine-set-operator-78cbb6b69f-hr5wj\" (UID: \"7dcad5af-901c-431f-b7b9-a98b4e667ee8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.447888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh4q4\" (UniqueName: \"kubernetes.io/projected/281b11d6-f406-4609-a432-5a796b2304ca-kube-api-access-kh4q4\") pod \"service-ca-9c57cc56f-zlmtd\" (UID: \"281b11d6-f406-4609-a432-5a796b2304ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.454891 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.456297 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wm56d"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.459530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-87qfj"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.464320 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.465214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvmxk\" (UniqueName: \"kubernetes.io/projected/918f0b1a-498d-4d82-a3e4-e0fcfa56ed48-kube-api-access-vvmxk\") pod \"openshift-config-operator-7777fb866f-rv274\" (UID: \"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.467138 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.467387 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.967367321 +0000 UTC m=+152.757392891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.467485 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.467903 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:41.967896101 +0000 UTC m=+152.757921681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.471437 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.483771 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwdjb\" (UniqueName: \"kubernetes.io/projected/8bce9738-a274-4c30-913e-37ea8663bee6-kube-api-access-nwdjb\") pod \"multus-admission-controller-857f4d67dd-t69dw\" (UID: \"8bce9738-a274-4c30-913e-37ea8663bee6\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.491057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.505548 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.512344 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.518610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjb7s\" (UniqueName: \"kubernetes.io/projected/51938ce0-b08d-456a-b711-b3de25465a9d-kube-api-access-kjb7s\") pod \"dns-operator-744455d44c-b2g98\" (UID: \"51938ce0-b08d-456a-b711-b3de25465a9d\") " pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.519916 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.524880 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.545597 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbz42\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.563260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt5t2\" (UniqueName: \"kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2\") pod \"marketplace-operator-79b997595-78t58\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.568511 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.568980 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.068963393 +0000 UTC m=+152.858988973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.582541 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.595865 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.603184 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.609510 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.613470 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.614722 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.622880 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.623943 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dkdk\" (UniqueName: \"kubernetes.io/projected/75c7b719-ca97-4dbe-8e30-3b3ac2160e5e-kube-api-access-2dkdk\") pod \"machine-config-controller-84d6567774-k4pcm\" (UID: \"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.626428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr648\" (UniqueName: \"kubernetes.io/projected/4a05a51e-c77e-4272-91fa-290cbec0f3d3-kube-api-access-lr648\") pod \"ingress-canary-4wcbb\" (UID: \"4a05a51e-c77e-4272-91fa-290cbec0f3d3\") " pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.631585 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.641069 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b2a364f-6a58-442e-9dc8-3807d92943e3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-74lh8\" (UID: \"4b2a364f-6a58-442e-9dc8-3807d92943e3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.652262 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.658883 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.661581 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kf8m\" (UniqueName: \"kubernetes.io/projected/2650c70c-4fa3-4fd3-8c76-53c7b2d33148-kube-api-access-4kf8m\") pod \"etcd-operator-b45778765-8zkhj\" (UID: \"2650c70c-4fa3-4fd3-8c76-53c7b2d33148\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.674809 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.676143 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.676489 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.176473877 +0000 UTC m=+152.966499467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.685262 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktcrq\" (UniqueName: \"kubernetes.io/projected/3c706c75-f68a-4fc7-9cc4-1de179bc7e8b-kube-api-access-ktcrq\") pod \"machine-config-operator-74547568cd-c87pf\" (UID: \"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.688810 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.692213 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" event={"ID":"14cad03b-9a44-4f7f-a3a1-ae8b970fd993","Type":"ContainerStarted","Data":"40d16538bdc6e303072d246235af03345f75d4d8ac1cb4f477dab8172b03483a"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.703835 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.708439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt7tp\" (UniqueName: \"kubernetes.io/projected/868aae56-7e7a-4af7-8c31-b97448d5c81f-kube-api-access-qt7tp\") pod \"kube-storage-version-migrator-operator-b67b599dd-74vq4\" (UID: \"868aae56-7e7a-4af7-8c31-b97448d5c81f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.709772 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wm56d" event={"ID":"595af3bc-3f4a-4214-a46b-4ea6b57136e2","Type":"ContainerStarted","Data":"b5a73a4a857c49788b386d79942f8ecea11a70d5f020fa6b08c1bf29fbe50073"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.709812 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wm56d" event={"ID":"595af3bc-3f4a-4214-a46b-4ea6b57136e2","Type":"ContainerStarted","Data":"c251000ef6e1c388a2516ddae10514296ce6e49f41fd06ac6bd453cc2f92d3d3"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.709891 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.724395 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.725105 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.730453 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvsjx\" (UniqueName: \"kubernetes.io/projected/41f81572-1755-48a1-b822-b13798b463a4-kube-api-access-zvsjx\") pod \"ingress-operator-5b745b69d9-npjhd\" (UID: \"41f81572-1755-48a1-b822-b13798b463a4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.736012 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.740696 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9pfhc" event={"ID":"3bd79c86-9633-4dcb-bc5f-800abe66a513","Type":"ContainerStarted","Data":"e4680a0326b34afd94e64214c48d37b7df0578bcf95e323f4a4ee8eb7d41e211"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.740772 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9pfhc" event={"ID":"3bd79c86-9633-4dcb-bc5f-800abe66a513","Type":"ContainerStarted","Data":"25e5e33db47de727541a0d30b1c033e1147dd8397ab5e36d1bf42dad2653744c"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.741571 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a4b702f6-d926-4d3e-b509-60facd64ed7c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-62rm5\" (UID: \"a4b702f6-d926-4d3e-b509-60facd64ed7c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.741626 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.743052 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4wcbb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.745783 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-9pfhc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.745816 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9pfhc" podUID="3bd79c86-9633-4dcb-bc5f-800abe66a513" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.748534 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" event={"ID":"057c0c57-faf9-439b-b8a1-f5f4a659f0d8","Type":"ContainerStarted","Data":"c6a029943888baf3020b1d4418c09bf1023ddbed2d8f29bfadb770d2122be221"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.748570 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" event={"ID":"057c0c57-faf9-439b-b8a1-f5f4a659f0d8","Type":"ContainerStarted","Data":"818cd34b6c84f3bc814730a33c627d770e7343fcd5ce5d2ad208cf9c586aec85"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.754203 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" event={"ID":"ee383275-f8d1-4dca-89e5-978bfa5ad4e7","Type":"ContainerStarted","Data":"59f782e12f02b8643880815320fe0ae3f12d2678a38483a3834c3640e01d611d"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.756457 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" event={"ID":"f085dab8-7067-4be2-844c-884ba9688549","Type":"ContainerStarted","Data":"a57f32c514e1650b88b8350af194292f9442a58d48fab66162a8fc32f4221d1a"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.756536 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" event={"ID":"f085dab8-7067-4be2-844c-884ba9688549","Type":"ContainerStarted","Data":"0275be3704c7974eb2ef4f929e8e5884f687801cb8daac17ec2e3fd6a0c98b26"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.757753 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" event={"ID":"61635015-0a25-4834-8f24-2ecee7ee9240","Type":"ContainerStarted","Data":"62a229507808807fff4c0fde3da474a39e351c345f43107e89b3063dba63c4dc"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.759222 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" event={"ID":"118c9ba6-fc48-4427-bf18-82c4ec482e20","Type":"ContainerStarted","Data":"ec1f67d8d484be85f5ae6cbb9ee597b532ebaea1256a80518fdc0c80b580ef35"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.760845 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wwr9w" event={"ID":"b3a7037e-5b9e-4440-a24c-aaca9364fe0f","Type":"ContainerStarted","Data":"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.760869 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wwr9w" event={"ID":"b3a7037e-5b9e-4440-a24c-aaca9364fe0f","Type":"ContainerStarted","Data":"bccda34975faf5a32dbbc1c66d61eb9efe43b1172525f89ec9001fd8e05e9558"} Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.765257 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmkc9\" (UniqueName: \"kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9\") pod \"collect-profiles-29328990-v4h88\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.776741 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.776936 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.276909305 +0000 UTC m=+153.066934885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.777355 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.777681 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.277649773 +0000 UTC m=+153.067675353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.784928 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flbj2\" (UniqueName: \"kubernetes.io/projected/a52e7f91-d84a-495d-a6ef-dcd0d0618990-kube-api-access-flbj2\") pod \"csi-hostpathplugin-zlrl6\" (UID: \"a52e7f91-d84a-495d-a6ef-dcd0d0618990\") " pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.809345 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h8dw\" (UniqueName: \"kubernetes.io/projected/2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120-kube-api-access-5h8dw\") pod \"service-ca-operator-777779d784-h9w9z\" (UID: \"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.814168 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.830469 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14c7313b-b022-44fc-bb7f-b35c3fea82de-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wc292\" (UID: \"14c7313b-b022-44fc-bb7f-b35c3fea82de\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.848047 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl2qg\" (UniqueName: \"kubernetes.io/projected/a87042d1-b2c4-44ac-9f0c-853b88375f53-kube-api-access-rl2qg\") pod \"openshift-controller-manager-operator-756b6f6bc6-xcw7d\" (UID: \"a87042d1-b2c4-44ac-9f0c-853b88375f53\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.866047 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87j2g\" (UniqueName: \"kubernetes.io/projected/0a46506a-7ebe-405b-8ff1-88a9122390e6-kube-api-access-87j2g\") pod \"dns-default-x9rxz\" (UID: \"0a46506a-7ebe-405b-8ff1-88a9122390e6\") " pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.869076 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.878325 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.878793 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.378774087 +0000 UTC m=+153.168799667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.884446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xgmg\" (UniqueName: \"kubernetes.io/projected/bd7241b3-d77b-458e-b4bf-77fbd9f28596-kube-api-access-6xgmg\") pod \"router-default-5444994796-c829n\" (UID: \"bd7241b3-d77b-458e-b4bf-77fbd9f28596\") " pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.905612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gppxq\" (UniqueName: \"kubernetes.io/projected/c76e2108-d346-4eeb-93b7-1180c2617371-kube-api-access-gppxq\") pod \"machine-config-server-6qqlb\" (UID: \"c76e2108-d346-4eeb-93b7-1180c2617371\") " pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.928084 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.937904 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" Oct 06 08:41:41 crc kubenswrapper[4989]: W1006 08:41:41.952879 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b4d3115_9287_41a4_9435_1d34590a5178.slice/crio-08c80de1a17b90f7d531aba4cfccfa6e6ee64b8d171eb3ed69b8d159bb333775 WatchSource:0}: Error finding container 08c80de1a17b90f7d531aba4cfccfa6e6ee64b8d171eb3ed69b8d159bb333775: Status 404 returned error can't find the container with id 08c80de1a17b90f7d531aba4cfccfa6e6ee64b8d171eb3ed69b8d159bb333775 Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.954196 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcq7v"] Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.966997 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.972343 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.982555 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" Oct 06 08:41:41 crc kubenswrapper[4989]: I1006 08:41:41.983407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:41 crc kubenswrapper[4989]: E1006 08:41:41.983750 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.483738896 +0000 UTC m=+153.273764476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:41.997508 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.025141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.035579 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.042711 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.052029 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6qqlb" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.059801 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.077114 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.084031 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.084234 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.584211015 +0000 UTC m=+153.374236595 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.084297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.085425 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.58541404 +0000 UTC m=+153.375439620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: W1006 08:41:42.178225 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa0ddc12_b6aa_4159_9945_042260a0595c.slice/crio-6d15a6e8ef02b4b664a39a72337bd2c33171bc51820ab51d5877a4ac6365a577 WatchSource:0}: Error finding container 6d15a6e8ef02b4b664a39a72337bd2c33171bc51820ab51d5877a4ac6365a577: Status 404 returned error can't find the container with id 6d15a6e8ef02b4b664a39a72337bd2c33171bc51820ab51d5877a4ac6365a577 Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.184838 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.185045 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.685019677 +0000 UTC m=+153.475045257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.185958 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.186446 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.686396749 +0000 UTC m=+153.476422319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.291403 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.291544 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.791522264 +0000 UTC m=+153.581547844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.291902 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.292272 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.792255111 +0000 UTC m=+153.582280691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: W1006 08:41:42.337632 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd7241b3_d77b_458e_b4bf_77fbd9f28596.slice/crio-68bf3614da32c9688f03d6c1ede133f7d908b99e13c0c4c99a0501af3d59b07e WatchSource:0}: Error finding container 68bf3614da32c9688f03d6c1ede133f7d908b99e13c0c4c99a0501af3d59b07e: Status 404 returned error can't find the container with id 68bf3614da32c9688f03d6c1ede133f7d908b99e13c0c4c99a0501af3d59b07e Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.388810 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wwr9w" podStartSLOduration=129.388795922 podStartE2EDuration="2m9.388795922s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:42.388120877 +0000 UTC m=+153.178146457" watchObservedRunningTime="2025-10-06 08:41:42.388795922 +0000 UTC m=+153.178821502" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.392743 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.393054 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.893036712 +0000 UTC m=+153.683062292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.406607 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-m7lpj"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.417271 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zlmtd"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.417313 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.494374 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.495163 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:42.995145493 +0000 UTC m=+153.785171073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.595065 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.595795 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.095776008 +0000 UTC m=+153.885801588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.696906 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.697252 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.197237655 +0000 UTC m=+153.987263235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.797930 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.798419 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.29838266 +0000 UTC m=+154.088408250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.799000 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.799572 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.299556924 +0000 UTC m=+154.089582504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.823915 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" event={"ID":"fb18713a-3d90-4486-aa7e-e93ff0718414","Type":"ContainerStarted","Data":"2d327f0fd39378fc7981480894414187f52b5a90b1baf0b26016e4eb5d067563"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.827469 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" event={"ID":"9ec320c3-776d-4197-9259-7feee562229a","Type":"ContainerStarted","Data":"d8a9b63ec679900d3a34ceeb457681ba19fcd463057947bbadeb78df08207ed7"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.828574 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" event={"ID":"fc8b7232-28ce-4453-a5cd-eddb6077f07d","Type":"ContainerStarted","Data":"30d0fc01f913b280f70e987f19546db6b3968f407481471863cf3e35dfaea319"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.828608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" event={"ID":"fc8b7232-28ce-4453-a5cd-eddb6077f07d","Type":"ContainerStarted","Data":"4da07a1b675c660a969a7f7243652d03b8764a9e7629d34d25ac9f651422fcc8"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.831331 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" event={"ID":"2b4d3115-9287-41a4-9435-1d34590a5178","Type":"ContainerStarted","Data":"f901633d593147837b285f836787e385dc23758abb8512b4b8b5d21caea4fbdb"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.831383 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" event={"ID":"2b4d3115-9287-41a4-9435-1d34590a5178","Type":"ContainerStarted","Data":"08c80de1a17b90f7d531aba4cfccfa6e6ee64b8d171eb3ed69b8d159bb333775"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.836531 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" event={"ID":"aa0ddc12-b6aa-4159-9945-042260a0595c","Type":"ContainerStarted","Data":"6d15a6e8ef02b4b664a39a72337bd2c33171bc51820ab51d5877a4ac6365a577"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.844309 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6qqlb" event={"ID":"c76e2108-d346-4eeb-93b7-1180c2617371","Type":"ContainerStarted","Data":"307a45e2d7f3d316c6923e034ec845532bce79c9db97d78a98afb2180f8b145f"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.858733 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" event={"ID":"057c0c57-faf9-439b-b8a1-f5f4a659f0d8","Type":"ContainerStarted","Data":"4c361e33546b553924e8c9afebce6983d47b31a6c8cdaccaed6fb91d7603fbe6"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.866968 4989 generic.go:334] "Generic (PLEG): container finished" podID="ee383275-f8d1-4dca-89e5-978bfa5ad4e7" containerID="b108e2e3208d7f0c80e5b49a4918bb19fc460b05e39d0be6e9fc8d4ba00b5ac0" exitCode=0 Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.867037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" event={"ID":"ee383275-f8d1-4dca-89e5-978bfa5ad4e7","Type":"ContainerDied","Data":"b108e2e3208d7f0c80e5b49a4918bb19fc460b05e39d0be6e9fc8d4ba00b5ac0"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.868850 4989 generic.go:334] "Generic (PLEG): container finished" podID="61635015-0a25-4834-8f24-2ecee7ee9240" containerID="fab4091b353186ab0db1170ac8c22429c737c1c15dd0eb84e964d51cc1a1d10d" exitCode=0 Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.868907 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" event={"ID":"61635015-0a25-4834-8f24-2ecee7ee9240","Type":"ContainerDied","Data":"fab4091b353186ab0db1170ac8c22429c737c1c15dd0eb84e964d51cc1a1d10d"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.879353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" event={"ID":"14cad03b-9a44-4f7f-a3a1-ae8b970fd993","Type":"ContainerStarted","Data":"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.880324 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.898342 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" event={"ID":"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff","Type":"ContainerStarted","Data":"4376618079be8ff6d2e0585cf190f9aa9133cde1ef3e86fd36bae626ac43dbc1"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.899448 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.899773 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:42 crc kubenswrapper[4989]: E1006 08:41:42.901463 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.401444316 +0000 UTC m=+154.191469896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.903677 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69dw"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.903714 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8zkhj"] Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.930395 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r96zs" podStartSLOduration=129.930370657 podStartE2EDuration="2m9.930370657s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:42.922549682 +0000 UTC m=+153.712575272" watchObservedRunningTime="2025-10-06 08:41:42.930370657 +0000 UTC m=+153.720396237" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.961647 4989 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ns27q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.973009 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" podUID="bf7a30ae-66ce-49b6-b59b-cdf42a8662ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.976989 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" event={"ID":"c8c4448a-914d-4e2c-9ad7-71a2d0444669","Type":"ContainerStarted","Data":"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.977038 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.977048 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" event={"ID":"c8c4448a-914d-4e2c-9ad7-71a2d0444669","Type":"ContainerStarted","Data":"6b896d6433e032207102d589e22e13a354109859756280afe16aa745e631a9b7"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.979974 4989 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ww9xn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.980023 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.998735 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" event={"ID":"92c1ecab-13ee-4b81-9454-ad02e7071592","Type":"ContainerStarted","Data":"0adb4833e45e7f69ef868e0a437fd0a52fb0f54a84cdf1344628496a649f9e3b"} Oct 06 08:41:42 crc kubenswrapper[4989]: I1006 08:41:42.998816 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" event={"ID":"92c1ecab-13ee-4b81-9454-ad02e7071592","Type":"ContainerStarted","Data":"ced2e679d5f530fba0e43541c9e607846812c518ad83f84e9ed35ac7f70a7849"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.001387 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.001737 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.501721908 +0000 UTC m=+154.291747488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.009906 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" event={"ID":"281b11d6-f406-4609-a432-5a796b2304ca","Type":"ContainerStarted","Data":"bca248ab8a20064f5f37546853fefaa1cc5e37bf02a6015df9b40e9384e0afab"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.021103 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-c829n" event={"ID":"bd7241b3-d77b-458e-b4bf-77fbd9f28596","Type":"ContainerStarted","Data":"68bf3614da32c9688f03d6c1ede133f7d908b99e13c0c4c99a0501af3d59b07e"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.043717 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" event={"ID":"97d62f70-c96b-4468-bc7f-c110115a03b6","Type":"ContainerStarted","Data":"6e7a50fd5c9051efd09ee2577289881958522ea5c1f7258113f29fa02171e8c1"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.043765 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" event={"ID":"97d62f70-c96b-4468-bc7f-c110115a03b6","Type":"ContainerStarted","Data":"db2669d6e3413dbddfdc9ec3143e9799738a51c2fe46bc59d3bcbbf70b222b9f"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.045078 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.048823 4989 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2cwvl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.049175 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" podUID="97d62f70-c96b-4468-bc7f-c110115a03b6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.068285 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" event={"ID":"118c9ba6-fc48-4427-bf18-82c4ec482e20","Type":"ContainerStarted","Data":"ddc75fb2a12009de2cd35c712af77986618bea488afc3da34f3e5e5c1583e422"} Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.070489 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.072251 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-9pfhc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.072285 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9pfhc" podUID="3bd79c86-9633-4dcb-bc5f-800abe66a513" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.074688 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.077084 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b2g98"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.083315 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.102165 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.103084 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.60305753 +0000 UTC m=+154.393083110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.103239 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.106064 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.606053963 +0000 UTC m=+154.396079543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.207676 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.208216 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.708169134 +0000 UTC m=+154.498194714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.260593 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.270850 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rv274"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.285218 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.296033 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4wcbb"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.305270 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.305330 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.313148 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.320821 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm"] Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.360709 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.859646507 +0000 UTC m=+154.649672087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.374528 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.377759 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.421112 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.421536 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:43.921515581 +0000 UTC m=+154.711541161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.501600 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.506735 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zlrl6"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.522208 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.523489 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.023474466 +0000 UTC m=+154.813500046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: W1006 08:41:43.538482 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dcad5af_901c_431f_b7b9_a98b4e667ee8.slice/crio-76590cb2c260722cb270d5c4079d7a4391e364fd766a25ace12fd94aa07cd258 WatchSource:0}: Error finding container 76590cb2c260722cb270d5c4079d7a4391e364fd766a25ace12fd94aa07cd258: Status 404 returned error can't find the container with id 76590cb2c260722cb270d5c4079d7a4391e364fd766a25ace12fd94aa07cd258 Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.541027 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.547379 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9pfhc" podStartSLOduration=130.547362997 podStartE2EDuration="2m10.547362997s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.540948345 +0000 UTC m=+154.330973925" watchObservedRunningTime="2025-10-06 08:41:43.547362997 +0000 UTC m=+154.337388577" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.547994 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.574513 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.580314 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x9rxz"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.584723 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d"] Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.627253 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.627512 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.127496039 +0000 UTC m=+154.917521619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: W1006 08:41:43.681041 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25ae2464_66b2_4986_beef_382caadb447e.slice/crio-e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680 WatchSource:0}: Error finding container e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680: Status 404 returned error can't find the container with id e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680 Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.700327 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" podStartSLOduration=129.700309275 podStartE2EDuration="2m9.700309275s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.671319252 +0000 UTC m=+154.461344832" watchObservedRunningTime="2025-10-06 08:41:43.700309275 +0000 UTC m=+154.490334855" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.728238 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.728567 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.228556231 +0000 UTC m=+155.018581811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.772624 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wm56d" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.786044 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" podStartSLOduration=129.786024368 podStartE2EDuration="2m9.786024368s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.744015424 +0000 UTC m=+154.534041004" watchObservedRunningTime="2025-10-06 08:41:43.786024368 +0000 UTC m=+154.576049948" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.852295 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.852851 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.352831698 +0000 UTC m=+155.142857278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.880818 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" podStartSLOduration=129.880803252 podStartE2EDuration="2m9.880803252s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.879552015 +0000 UTC m=+154.669577595" watchObservedRunningTime="2025-10-06 08:41:43.880803252 +0000 UTC m=+154.670828832" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.881317 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mtfp9" podStartSLOduration=130.881313102 podStartE2EDuration="2m10.881313102s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.821443704 +0000 UTC m=+154.611469284" watchObservedRunningTime="2025-10-06 08:41:43.881313102 +0000 UTC m=+154.671338682" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.908266 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcq7v" podStartSLOduration=130.908249758 podStartE2EDuration="2m10.908249758s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:43.907398845 +0000 UTC m=+154.697424415" watchObservedRunningTime="2025-10-06 08:41:43.908249758 +0000 UTC m=+154.698275338" Oct 06 08:41:43 crc kubenswrapper[4989]: I1006 08:41:43.955101 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:43 crc kubenswrapper[4989]: E1006 08:41:43.955413 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.455399376 +0000 UTC m=+155.245424956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.057252 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.057412 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.557388412 +0000 UTC m=+155.347413992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.057505 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.057954 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.557939153 +0000 UTC m=+155.347964733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.086896 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" event={"ID":"281b11d6-f406-4609-a432-5a796b2304ca","Type":"ContainerStarted","Data":"07810e3d1d994809b29993c0780230c67feeab988ef920b64b36fa2d7f0950bb"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.088998 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9rxz" event={"ID":"0a46506a-7ebe-405b-8ff1-88a9122390e6","Type":"ContainerStarted","Data":"3cb5a78cdf6df6d76097a7ff7ac72a3bb2ea1c2ea83e7261070e1e1699177603"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.090686 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" event={"ID":"868aae56-7e7a-4af7-8c31-b97448d5c81f","Type":"ContainerStarted","Data":"6fce422d5acbdeb3a2549933d80ebd616e215385aafe88c01cc25bfccd4fee48"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.096949 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" event={"ID":"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b","Type":"ContainerStarted","Data":"46473a97d585638f3ef3db2e649a54f13b9bb2ba5945d441ff30b84115274ec6"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.099679 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" event={"ID":"72aca57c-867e-460f-b3c9-8d0a2bc275ca","Type":"ContainerStarted","Data":"8e2e48d6511f7e4de601386f2047103576ee9c5e1be68a1e1c5cbe5f5b832a6a"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.099711 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" event={"ID":"72aca57c-867e-460f-b3c9-8d0a2bc275ca","Type":"ContainerStarted","Data":"2dc3b2ce0105d886d6e9872ae1678e07dcdfbfeb6a5f1b0d7a27365be58311be"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.101028 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" event={"ID":"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e","Type":"ContainerStarted","Data":"ab2f0f2ced1a0f675dcd208c80deaac1c761b0117a9e02c23f192f1496235a2e"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.118527 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wm56d" podStartSLOduration=131.118496957 podStartE2EDuration="2m11.118496957s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.077330694 +0000 UTC m=+154.867356294" watchObservedRunningTime="2025-10-06 08:41:44.118496957 +0000 UTC m=+154.908522547" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.130894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" event={"ID":"bf7a30ae-66ce-49b6-b59b-cdf42a8662ff","Type":"ContainerStarted","Data":"3394fd60fe80b58a78ec46cd7db31bd92c50fabce7a5dcd3b19cbafad3efab21"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.131977 4989 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ns27q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.132024 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" podUID="bf7a30ae-66ce-49b6-b59b-cdf42a8662ff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.135050 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" event={"ID":"a52e7f91-d84a-495d-a6ef-dcd0d0618990","Type":"ContainerStarted","Data":"c92474156fb68217bc8e07c43aa436303aad40e64d630d0b46f050e3f977ef1b"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.142545 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" event={"ID":"7dcad5af-901c-431f-b7b9-a98b4e667ee8","Type":"ContainerStarted","Data":"76590cb2c260722cb270d5c4079d7a4391e364fd766a25ace12fd94aa07cd258"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.149406 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-c829n" event={"ID":"bd7241b3-d77b-458e-b4bf-77fbd9f28596","Type":"ContainerStarted","Data":"f4c7e77d0e1c91991599ba79cb09413b1dc95c797b98b2308ef5e77cb4d04a11"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.160232 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.161623 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.661600513 +0000 UTC m=+155.451626093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.184516 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" event={"ID":"8bce9738-a274-4c30-913e-37ea8663bee6","Type":"ContainerStarted","Data":"10879535bdeb788622037312333a22edbc7f4ec1b43ff97100dc98e2b55062e4"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.184562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" event={"ID":"8bce9738-a274-4c30-913e-37ea8663bee6","Type":"ContainerStarted","Data":"596e840edd960c890386e56b0dad8a56b5e71d05cc4bfa8fc79f04317478d870"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.192248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" event={"ID":"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48","Type":"ContainerStarted","Data":"dd21d9b4dbe9b04f7575ee7d6b9813cbd137a271b3f2195f0295bcc71a0d4628"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.219847 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" event={"ID":"51938ce0-b08d-456a-b711-b3de25465a9d","Type":"ContainerStarted","Data":"a4746ee83121b9fa47b1b42e0ba05606402d2213f423d4fcf3b7ab945f322a6b"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.219889 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" event={"ID":"51938ce0-b08d-456a-b711-b3de25465a9d","Type":"ContainerStarted","Data":"9c21d9973ff388215dcfc87f040d26bddec934d11d59f8c6da0644a672281b2e"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.263112 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.264893 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.764877328 +0000 UTC m=+155.554902908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.289851 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" event={"ID":"ee383275-f8d1-4dca-89e5-978bfa5ad4e7","Type":"ContainerStarted","Data":"75b4ac71eec79f0998368264f2b7518df0dc95eb2a5c36c36311d9277bc9185e"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.320753 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" event={"ID":"92c1ecab-13ee-4b81-9454-ad02e7071592","Type":"ContainerStarted","Data":"af861540a8864438244fadeb3a0df5c612d4b12877397313b3391c0d58a14e0b"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.323428 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" event={"ID":"9ec320c3-776d-4197-9259-7feee562229a","Type":"ContainerStarted","Data":"9cf84782aa5fae0d264b111ae53866c65b693dfa8d9ed4bf0790058ac19721b1"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.340946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6qqlb" event={"ID":"c76e2108-d346-4eeb-93b7-1180c2617371","Type":"ContainerStarted","Data":"0c8bcdae8843add8b7037322f28f4ccf236db834d2bc29c7d7d09aa43cd0aa23"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.365145 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.368368 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.868334169 +0000 UTC m=+155.658359749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.373795 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4wcbb" event={"ID":"4a05a51e-c77e-4272-91fa-290cbec0f3d3","Type":"ContainerStarted","Data":"96325a5c63dcf7de50d9bf60bff6209fd6ac13a20c9264f5b1e0d83b2da34ac8"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.380913 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" event={"ID":"a87042d1-b2c4-44ac-9f0c-853b88375f53","Type":"ContainerStarted","Data":"10450be1db7244ae3421b6267e1b3b5e7eb0e60a38fdcc541f27afd4ff3fe40d"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.412273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" event={"ID":"118c9ba6-fc48-4427-bf18-82c4ec482e20","Type":"ContainerStarted","Data":"68f2324871d823037f114a602d99a406e83898a6389ae094556fda546401dc82"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.445183 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" event={"ID":"61635015-0a25-4834-8f24-2ecee7ee9240","Type":"ContainerStarted","Data":"17afa4c1143759c624f8d3792963a6647a77855c1b85ccc40f30f3234ee19533"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.465074 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" event={"ID":"4b2a364f-6a58-442e-9dc8-3807d92943e3","Type":"ContainerStarted","Data":"f18f1f877c3394a0f4caa603e78a82a344337aeba32dbb565eedb3dfe0626dc4"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.465552 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" event={"ID":"4b2a364f-6a58-442e-9dc8-3807d92943e3","Type":"ContainerStarted","Data":"3a43d82edbcec4f85047631e5b7872acc218c153a4b9fa26a759a2a483095589"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.471914 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.475204 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:44.975189299 +0000 UTC m=+155.765214879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.504566 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" event={"ID":"25ae2464-66b2-4986-beef-382caadb447e","Type":"ContainerStarted","Data":"e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.539343 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" podStartSLOduration=130.539327498 podStartE2EDuration="2m10.539327498s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.507728907 +0000 UTC m=+155.297754487" watchObservedRunningTime="2025-10-06 08:41:44.539327498 +0000 UTC m=+155.329353078" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.578060 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" event={"ID":"2650c70c-4fa3-4fd3-8c76-53c7b2d33148","Type":"ContainerStarted","Data":"3cbf16f94009cecf7554b5d67b7feda9a9d19bf1b0adf5e2c882f39ce6e4c7b5"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.578114 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" event={"ID":"2650c70c-4fa3-4fd3-8c76-53c7b2d33148","Type":"ContainerStarted","Data":"d35abb17845126fc5fe1138aa9172214a768cd2c6bef65d79ce4b642d3ad98ee"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.582430 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.584157 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.084130878 +0000 UTC m=+155.874156458 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.598675 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" event={"ID":"14c7313b-b022-44fc-bb7f-b35c3fea82de","Type":"ContainerStarted","Data":"014065590afd4511002ace6d786b3181b5e59148ae452e931a42bada5a7b2bef"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.609899 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" event={"ID":"aa0ddc12-b6aa-4159-9945-042260a0595c","Type":"ContainerStarted","Data":"820d3dc7e9f8125e51ae4bd11c0c9e64dd015c043c9d2315f764d70e0204292f"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.634292 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" podStartSLOduration=131.634271389 podStartE2EDuration="2m11.634271389s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.632366327 +0000 UTC m=+155.422391907" watchObservedRunningTime="2025-10-06 08:41:44.634271389 +0000 UTC m=+155.424296969" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.640542 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" event={"ID":"41f81572-1755-48a1-b822-b13798b463a4","Type":"ContainerStarted","Data":"58be74a63a0e89ae13ca1124aea8abef02e1715924603e8a9a32f2974c463397"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.643308 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" event={"ID":"6d858cfb-196c-40b5-8b67-de2529a4d3c4","Type":"ContainerStarted","Data":"f28c3ec16d494d8d414ef056cee3ceab16668b3abd2d59a2354517bb3a64c187"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.644242 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.664933 4989 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s2rlf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.665021 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" podUID="6d858cfb-196c-40b5-8b67-de2529a4d3c4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.678185 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" event={"ID":"fb18713a-3d90-4486-aa7e-e93ff0718414","Type":"ContainerStarted","Data":"6f5821254c80639f2d6c7c794ee6157ce2fabe5e2afb83602f7823008662b330"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.678793 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.685551 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.686532 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6t7bd" podStartSLOduration=131.68651773 podStartE2EDuration="2m11.68651773s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.683731224 +0000 UTC m=+155.473756804" watchObservedRunningTime="2025-10-06 08:41:44.68651773 +0000 UTC m=+155.476543310" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.686867 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" event={"ID":"a4b702f6-d926-4d3e-b509-60facd64ed7c","Type":"ContainerStarted","Data":"8784c5a3a336102b95f0dbfae8d734d1a931a9dc03e1d8f654c00ff7b3173d49"} Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.688057 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.188044877 +0000 UTC m=+155.978070457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.700476 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" event={"ID":"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120","Type":"ContainerStarted","Data":"289b7244a0066dc9dc39a0dfa505fd00a3b5696b38635e3d141a768dfa357452"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.700524 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" event={"ID":"2c27ed51-7ca5-4c4b-8f1e-d8ef4133b120","Type":"ContainerStarted","Data":"abc0ed3aa0aa95717ae354670122f53c77c21392934a67361253ec158b0ec555"} Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.710736 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-9pfhc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.710801 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9pfhc" podUID="3bd79c86-9633-4dcb-bc5f-800abe66a513" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.715730 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.717576 4989 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-78t58 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.717639 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.721882 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.736850 4989 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vhcsz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.736893 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.741201 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6qqlb" podStartSLOduration=5.741168481 podStartE2EDuration="5.741168481s" podCreationTimestamp="2025-10-06 08:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.727589889 +0000 UTC m=+155.517615479" watchObservedRunningTime="2025-10-06 08:41:44.741168481 +0000 UTC m=+155.531194071" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.787140 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.791743 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.291714387 +0000 UTC m=+156.081739967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.810143 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" podStartSLOduration=130.810128061 podStartE2EDuration="2m10.810128061s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.808092505 +0000 UTC m=+155.598118085" watchObservedRunningTime="2025-10-06 08:41:44.810128061 +0000 UTC m=+155.600153631" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.811134 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zlmtd" podStartSLOduration=130.811128119 podStartE2EDuration="2m10.811128119s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.776721932 +0000 UTC m=+155.566747512" watchObservedRunningTime="2025-10-06 08:41:44.811128119 +0000 UTC m=+155.601153699" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.891626 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-c829n" podStartSLOduration=130.891606704 podStartE2EDuration="2m10.891606704s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.852334163 +0000 UTC m=+155.642359733" watchObservedRunningTime="2025-10-06 08:41:44.891606704 +0000 UTC m=+155.681632284" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.896271 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.896578 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.396565211 +0000 UTC m=+156.186590791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.952335 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8zkhj" podStartSLOduration=130.952315804 podStartE2EDuration="2m10.952315804s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.924429912 +0000 UTC m=+155.714455492" watchObservedRunningTime="2025-10-06 08:41:44.952315804 +0000 UTC m=+155.742341374" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.992823 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4wcbb" podStartSLOduration=6.992806971 podStartE2EDuration="6.992806971s" podCreationTimestamp="2025-10-06 08:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:44.992170837 +0000 UTC m=+155.782196417" watchObservedRunningTime="2025-10-06 08:41:44.992806971 +0000 UTC m=+155.782832551" Oct 06 08:41:44 crc kubenswrapper[4989]: I1006 08:41:44.998095 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:44 crc kubenswrapper[4989]: E1006 08:41:44.998421 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.498403932 +0000 UTC m=+156.288429502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.026299 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" podStartSLOduration=131.026272832 podStartE2EDuration="2m11.026272832s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.02516346 +0000 UTC m=+155.815189050" watchObservedRunningTime="2025-10-06 08:41:45.026272832 +0000 UTC m=+155.816298402" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.027059 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.046079 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xql97" podStartSLOduration=131.046059798 podStartE2EDuration="2m11.046059798s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.04425229 +0000 UTC m=+155.834277870" watchObservedRunningTime="2025-10-06 08:41:45.046059798 +0000 UTC m=+155.836085378" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.052642 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:45 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:45 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:45 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.052752 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.093993 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" podStartSLOduration=131.093961405 podStartE2EDuration="2m11.093961405s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.091027094 +0000 UTC m=+155.881052674" watchObservedRunningTime="2025-10-06 08:41:45.093961405 +0000 UTC m=+155.883986975" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.101438 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.101779 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.601763989 +0000 UTC m=+156.391789569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.145146 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fx7cq" podStartSLOduration=131.145107384 podStartE2EDuration="2m11.145107384s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.131978199 +0000 UTC m=+155.922003779" watchObservedRunningTime="2025-10-06 08:41:45.145107384 +0000 UTC m=+155.935132974" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.180367 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-74lh8" podStartSLOduration=131.180349403 podStartE2EDuration="2m11.180349403s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.178000365 +0000 UTC m=+155.968025945" watchObservedRunningTime="2025-10-06 08:41:45.180349403 +0000 UTC m=+155.970374983" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.204103 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.204410 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.70439588 +0000 UTC m=+156.494421460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.253240 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" podStartSLOduration=131.253216771 podStartE2EDuration="2m11.253216771s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.212197044 +0000 UTC m=+156.002222614" watchObservedRunningTime="2025-10-06 08:41:45.253216771 +0000 UTC m=+156.043242351" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.256396 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" podStartSLOduration=131.256385511 podStartE2EDuration="2m11.256385511s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.240161879 +0000 UTC m=+156.030187449" watchObservedRunningTime="2025-10-06 08:41:45.256385511 +0000 UTC m=+156.046411091" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.305206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.305566 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.805550525 +0000 UTC m=+156.595576105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.406310 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.406594 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:45.906553744 +0000 UTC m=+156.696579324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.507797 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.508378 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.008346213 +0000 UTC m=+156.798371793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.608773 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.609346 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.109330372 +0000 UTC m=+156.899355952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.639036 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2cwvl" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.673938 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h9w9z" podStartSLOduration=131.673920658 podStartE2EDuration="2m11.673920658s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.35394035 +0000 UTC m=+156.143965930" watchObservedRunningTime="2025-10-06 08:41:45.673920658 +0000 UTC m=+156.463946238" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.707753 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" event={"ID":"14c7313b-b022-44fc-bb7f-b35c3fea82de","Type":"ContainerStarted","Data":"8d02f5e0ea6d3687ee6c8e4946ed5191acc3d48191344e20b949429d59fb660f"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.709838 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.710163 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.210153414 +0000 UTC m=+157.000178994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.712865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" event={"ID":"8bce9738-a274-4c30-913e-37ea8663bee6","Type":"ContainerStarted","Data":"4ec179d933c66e6e8ea0fdc38f622f287783228c80b44ba82b5bddfcd7dd1725"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.714631 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" event={"ID":"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e","Type":"ContainerStarted","Data":"8d1e5fbc395ff16392e9f37df7843efafdcb06d1ec0ba511b20d569aaa4875cc"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.714674 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" event={"ID":"75c7b719-ca97-4dbe-8e30-3b3ac2160e5e","Type":"ContainerStarted","Data":"e61c2314c0f134d77644b3dfda702610cc31c29a6ee51ed4442c885846152c10"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.716248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" event={"ID":"6d858cfb-196c-40b5-8b67-de2529a4d3c4","Type":"ContainerStarted","Data":"dd1dd993498af8cda209f176ea0fcdfa6f4c1a39e7f1c24af54ce18e107a0bb0"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.717218 4989 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-s2rlf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.717362 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" podUID="6d858cfb-196c-40b5-8b67-de2529a4d3c4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.717730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4wcbb" event={"ID":"4a05a51e-c77e-4272-91fa-290cbec0f3d3","Type":"ContainerStarted","Data":"209a82e702eab80ad607b93f9a5e78d134699ee82e7ae99b49c3f9133dd5bc88"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.719208 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" event={"ID":"25ae2464-66b2-4986-beef-382caadb447e","Type":"ContainerStarted","Data":"4d953e5712bd3c81aab1f7118d30d60f3d549a967fbebf044b5f14a3ce7546c3"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.721035 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" event={"ID":"868aae56-7e7a-4af7-8c31-b97448d5c81f","Type":"ContainerStarted","Data":"baaf6498a048463d59bd691e9877e35a6651255c4c7dead31fea37400ec8fcd9"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.722931 4989 generic.go:334] "Generic (PLEG): container finished" podID="918f0b1a-498d-4d82-a3e4-e0fcfa56ed48" containerID="f1eac2b10c620d08e9f5fa9363ae6e79f4c8230c5ea744d7bed1e990da1267e5" exitCode=0 Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.723039 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" event={"ID":"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48","Type":"ContainerDied","Data":"f1eac2b10c620d08e9f5fa9363ae6e79f4c8230c5ea744d7bed1e990da1267e5"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.724585 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" event={"ID":"72aca57c-867e-460f-b3c9-8d0a2bc275ca","Type":"ContainerStarted","Data":"536616ab689897263eb2075afae9b18f124c36a50cfb62291ecf81e402341078"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.725074 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.726231 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" event={"ID":"7dcad5af-901c-431f-b7b9-a98b4e667ee8","Type":"ContainerStarted","Data":"e896f9f42f68f38d0e99b88351d9a9078d6e1a10e265383b1ba76dbe6a8b3510"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.728841 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" event={"ID":"a87042d1-b2c4-44ac-9f0c-853b88375f53","Type":"ContainerStarted","Data":"13337a4b454c50d10a3501b83b99728cc757303015c74da4cc857c3f5de10df5"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.731539 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" event={"ID":"a52e7f91-d84a-495d-a6ef-dcd0d0618990","Type":"ContainerStarted","Data":"cd7d341ee8a40ab45a007ac3a144720b7d0976350d76fcd3284608d30bd08d5e"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.733507 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-m7lpj" event={"ID":"9ec320c3-776d-4197-9259-7feee562229a","Type":"ContainerStarted","Data":"365675953c224123e4c1998016e66ccec213e554fe383b6b87df09f5a1b3a900"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.735788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9rxz" event={"ID":"0a46506a-7ebe-405b-8ff1-88a9122390e6","Type":"ContainerStarted","Data":"5cc9322b56b201946bad74e7529aa55a1ece3b73e052c7e7b0e41108db2de7d5"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.735822 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x9rxz" event={"ID":"0a46506a-7ebe-405b-8ff1-88a9122390e6","Type":"ContainerStarted","Data":"1e328dab972014046cbae4946fdc8a8be062dfe494c5c5927016f0c465dce2a7"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.736420 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.745309 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" event={"ID":"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b","Type":"ContainerStarted","Data":"6254b3100033e1eb65b1c7af7ec67922d175c9914b7fb9897295fc79c8213bfa"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.745349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" event={"ID":"3c706c75-f68a-4fc7-9cc4-1de179bc7e8b","Type":"ContainerStarted","Data":"e5328aabe9b2fba7f679695875282620ce7d40fef17954bb04990f54a2157cda"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.748928 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" event={"ID":"41f81572-1755-48a1-b822-b13798b463a4","Type":"ContainerStarted","Data":"6f6816ceadd9ecd634dd92b819852fcb3e9b033f1685d60268f7c332923bf5f1"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.748957 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" event={"ID":"41f81572-1755-48a1-b822-b13798b463a4","Type":"ContainerStarted","Data":"47af625500963290c6172bbe85720703627b87a6c9dfd0d497bebf1daf0266bc"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.767771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" event={"ID":"51938ce0-b08d-456a-b711-b3de25465a9d","Type":"ContainerStarted","Data":"ae829608996f7e3de71e71c8f2e02c888ccafa0d8fb5aa680750ddc9fa0d7afa"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.784335 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-62rm5" event={"ID":"a4b702f6-d926-4d3e-b509-60facd64ed7c","Type":"ContainerStarted","Data":"5de6cc06018857667ec1d45fa5e9184f95d917aa1d0eb2de899d7c3bb43917af"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.804035 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wc292" podStartSLOduration=131.804019274 podStartE2EDuration="2m11.804019274s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.765042665 +0000 UTC m=+156.555068245" watchObservedRunningTime="2025-10-06 08:41:45.804019274 +0000 UTC m=+156.594044854" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.804482 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69dw" podStartSLOduration=131.804475102 podStartE2EDuration="2m11.804475102s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.802971605 +0000 UTC m=+156.592997185" watchObservedRunningTime="2025-10-06 08:41:45.804475102 +0000 UTC m=+156.594500682" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.810505 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.812867 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" event={"ID":"61635015-0a25-4834-8f24-2ecee7ee9240","Type":"ContainerStarted","Data":"355dc0b28ca0e7e5f2c3253b870286e1a29bdbfa7cae84942ec18b515073c299"} Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.818753 4989 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-78t58 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.819053 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.857992 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.357953329 +0000 UTC m=+157.147978909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.873012 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.873168 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ns27q" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.907566 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hr5wj" podStartSLOduration=131.907544909 podStartE2EDuration="2m11.907544909s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.873729074 +0000 UTC m=+156.663754674" watchObservedRunningTime="2025-10-06 08:41:45.907544909 +0000 UTC m=+156.697570479" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.942546 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:45 crc kubenswrapper[4989]: E1006 08:41:45.970470 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.470452991 +0000 UTC m=+157.260478571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.986389 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-c87pf" podStartSLOduration=131.986372432 podStartE2EDuration="2m11.986372432s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:45.894152904 +0000 UTC m=+156.684178484" watchObservedRunningTime="2025-10-06 08:41:45.986372432 +0000 UTC m=+156.776398012" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.992813 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:45 crc kubenswrapper[4989]: I1006 08:41:45.992863 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.009114 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.009536 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.017807 4989 patch_prober.go:28] interesting pod/apiserver-76f77b778f-87qfj container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.017895 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" podUID="61635015-0a25-4834-8f24-2ecee7ee9240" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.033184 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:46 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:46 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:46 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.033302 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.044101 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.044487 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.544469123 +0000 UTC m=+157.334494703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.054954 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x9rxz" podStartSLOduration=7.054937438 podStartE2EDuration="7.054937438s" podCreationTimestamp="2025-10-06 08:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.001983521 +0000 UTC m=+156.792009101" watchObservedRunningTime="2025-10-06 08:41:46.054937438 +0000 UTC m=+156.844963018" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.134161 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" podStartSLOduration=133.134130734 podStartE2EDuration="2m13.134130734s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.112736718 +0000 UTC m=+156.902762298" watchObservedRunningTime="2025-10-06 08:41:46.134130734 +0000 UTC m=+156.924156314" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.137582 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.146151 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.146620 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.646604615 +0000 UTC m=+157.436630195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.155233 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.174530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.178908 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.187136 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-74vq4" podStartSLOduration=132.187123233 podStartE2EDuration="2m12.187123233s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.183808148 +0000 UTC m=+156.973833728" watchObservedRunningTime="2025-10-06 08:41:46.187123233 +0000 UTC m=+156.977148813" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.243043 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-k4pcm" podStartSLOduration=132.243025141 podStartE2EDuration="2m12.243025141s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.215552045 +0000 UTC m=+157.005577625" watchObservedRunningTime="2025-10-06 08:41:46.243025141 +0000 UTC m=+157.033050711" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.246880 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.247261 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.247351 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.247429 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdwd4\" (UniqueName: \"kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.247601 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.747583223 +0000 UTC m=+157.537608803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.292325 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xcw7d" podStartSLOduration=132.29230919 podStartE2EDuration="2m12.29230919s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.279483056 +0000 UTC m=+157.069508636" watchObservedRunningTime="2025-10-06 08:41:46.29230919 +0000 UTC m=+157.082334770" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.293763 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.294604 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.303834 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.304905 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.315520 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" podStartSLOduration=132.315497535 podStartE2EDuration="2m12.315497535s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.314123623 +0000 UTC m=+157.104149203" watchObservedRunningTime="2025-10-06 08:41:46.315497535 +0000 UTC m=+157.105523105" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.331408 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349684 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349768 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349826 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349856 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349886 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl7g2\" (UniqueName: \"kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349917 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdwd4\" (UniqueName: \"kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.349943 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.350595 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.350706 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.850687152 +0000 UTC m=+157.640712732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.350941 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.390532 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-npjhd" podStartSLOduration=132.390514464 podStartE2EDuration="2m12.390514464s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.351387618 +0000 UTC m=+157.141413208" watchObservedRunningTime="2025-10-06 08:41:46.390514464 +0000 UTC m=+157.180540044" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.412211 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdwd4\" (UniqueName: \"kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4\") pod \"certified-operators-7xfhm\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.444367 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-b2g98" podStartSLOduration=132.444350644 podStartE2EDuration="2m12.444350644s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.443239752 +0000 UTC m=+157.233265332" watchObservedRunningTime="2025-10-06 08:41:46.444350644 +0000 UTC m=+157.234376224" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.451418 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.451733 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.451775 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl7g2\" (UniqueName: \"kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.451801 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.452302 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.452385 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:46.952364006 +0000 UTC m=+157.742389586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.452589 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.477957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.484354 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl7g2\" (UniqueName: \"kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2\") pod \"community-operators-tnvzq\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.559046 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.559404 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.059392153 +0000 UTC m=+157.849417733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.597407 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.600380 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.612551 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.618240 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.629211 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" podStartSLOduration=133.629184935 podStartE2EDuration="2m13.629184935s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.615495399 +0000 UTC m=+157.405520979" watchObservedRunningTime="2025-10-06 08:41:46.629184935 +0000 UTC m=+157.419210515" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.664266 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.664392 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.164368142 +0000 UTC m=+157.954393722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.665282 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.665342 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.665388 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp8wh\" (UniqueName: \"kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.665461 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.665857 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.165848758 +0000 UTC m=+157.955874338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.718056 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.719366 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.760051 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772236 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772386 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772430 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp8wh\" (UniqueName: \"kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772457 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb5sh\" (UniqueName: \"kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772501 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772529 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772554 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.772979 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.773048 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.27303227 +0000 UTC m=+158.063057840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.773527 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.814610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp8wh\" (UniqueName: \"kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh\") pod \"certified-operators-9rl5s\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.875583 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.875683 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb5sh\" (UniqueName: \"kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.875760 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.875796 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.876165 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.376147619 +0000 UTC m=+158.166173199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.877163 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.878160 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.887201 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" event={"ID":"918f0b1a-498d-4d82-a3e4-e0fcfa56ed48","Type":"ContainerStarted","Data":"00fd709c027e1f695102c1b5db98cf50995016d2e495bc559318f69cf897cc5b"} Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.892059 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.897472 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rrqq6" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.933800 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" podStartSLOduration=133.933785753 podStartE2EDuration="2m13.933785753s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:46.931086381 +0000 UTC m=+157.721111961" watchObservedRunningTime="2025-10-06 08:41:46.933785753 +0000 UTC m=+157.723811333" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.948942 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.949616 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-s2rlf" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.954610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb5sh\" (UniqueName: \"kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh\") pod \"community-operators-rqnb9\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:46 crc kubenswrapper[4989]: I1006 08:41:46.995186 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:46 crc kubenswrapper[4989]: E1006 08:41:46.996953 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.496938765 +0000 UTC m=+158.286964345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.032359 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:47 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:47 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:47 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.032410 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.042523 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.098735 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.099672 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.599633538 +0000 UTC m=+158.389659118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.201106 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.201384 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.701370005 +0000 UTC m=+158.491395585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.303296 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.303599 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.80358674 +0000 UTC m=+158.593612320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.371198 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.404546 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.404768 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.904753155 +0000 UTC m=+158.694778735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.405004 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.405253 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:47.905246804 +0000 UTC m=+158.695272384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.506247 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.506635 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.006617787 +0000 UTC m=+158.796643367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.610010 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.610383 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.11036748 +0000 UTC m=+158.900393060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.685931 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.711383 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.711781 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.211764124 +0000 UTC m=+159.001789704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.824954 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.825406 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.325389039 +0000 UTC m=+159.115414619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.916684 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.926403 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:47 crc kubenswrapper[4989]: E1006 08:41:47.926997 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.42697409 +0000 UTC m=+159.216999660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.959875 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerStarted","Data":"7a911560c3e865cba1f7875fbb1fc46fe7f621d571b166e361517971d51a95be"} Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.959919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerStarted","Data":"f98b19b68a80ef374e7dc538da5a485a1c72a8c797f3fba0f83a9df25d2ac34b"} Oct 06 08:41:47 crc kubenswrapper[4989]: I1006 08:41:47.981488 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.028309 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.028400 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" event={"ID":"a52e7f91-d84a-495d-a6ef-dcd0d0618990","Type":"ContainerStarted","Data":"83098ac44863acc9829d882f9d641d0487b39322c58476486c5f123cec416bc3"} Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.028611 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.528599623 +0000 UTC m=+159.318625203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.031134 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:48 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:48 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:48 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.031203 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.056450 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerStarted","Data":"e946f489117ae219551744bb14b97acd970d8a90911fe2730ad89770b5410be8"} Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.129662 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.130646 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.630624921 +0000 UTC m=+159.420650501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.231298 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.231913 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.73187641 +0000 UTC m=+159.521901990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.332271 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.332404 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.8323819 +0000 UTC m=+159.622407480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.333188 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.333553 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.833537534 +0000 UTC m=+159.623563114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.434774 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.434883 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.934864215 +0000 UTC m=+159.724889795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.435149 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.435408 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:48.935398815 +0000 UTC m=+159.725424395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.508764 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.510302 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.512913 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.531291 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.536336 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.536648 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.036613793 +0000 UTC m=+159.826639373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.638028 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8xzl\" (UniqueName: \"kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.638077 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.638111 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.638242 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.638533 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.138519325 +0000 UTC m=+159.928544905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.739071 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.739280 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.239251704 +0000 UTC m=+160.029277294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.740115 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.740490 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.24047779 +0000 UTC m=+160.030503370 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.740844 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8xzl\" (UniqueName: \"kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.741248 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.741938 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.741802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.742277 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.763721 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8xzl\" (UniqueName: \"kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl\") pod \"redhat-marketplace-4q4mm\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.843272 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.843408 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.343391632 +0000 UTC m=+160.133417212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.843811 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.844175 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.344157231 +0000 UTC m=+160.134182811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.845076 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.877146 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.878044 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.897408 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:41:48 crc kubenswrapper[4989]: I1006 08:41:48.944558 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:48 crc kubenswrapper[4989]: E1006 08:41:48.944961 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.444945462 +0000 UTC m=+160.234971042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.036072 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:49 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:49 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:49 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.036272 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.045866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.045927 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.045954 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.046018 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbwhf\" (UniqueName: \"kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.046275 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.546259563 +0000 UTC m=+160.336285143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.091059 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerID="973ad1862e4fbcc7bffb26582189156a7baa09337bd3ab034e2f5f420c94edfd" exitCode=0 Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.091800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerDied","Data":"973ad1862e4fbcc7bffb26582189156a7baa09337bd3ab034e2f5f420c94edfd"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.091829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerStarted","Data":"99894760e93a8d92cd98ff9b27cece2618eeee3116230f6e4e601a1183b3a646"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.099231 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.101492 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerID="091dd380f97b2731ca1b118960f0f724454c122308c32eb9f618abd2b9274e2d" exitCode=0 Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.101566 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerDied","Data":"091dd380f97b2731ca1b118960f0f724454c122308c32eb9f618abd2b9274e2d"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.101598 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerStarted","Data":"a7037daea8a158c5a7b7e1c4d7e9ee92a4caaf5c3191ddc39c7d19059a517c0c"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.118604 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" event={"ID":"a52e7f91-d84a-495d-a6ef-dcd0d0618990","Type":"ContainerStarted","Data":"e231a9ac981844440ec49262afd128f068cebbe246e279e56c0b2ff0d3b4e00f"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.118641 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" event={"ID":"a52e7f91-d84a-495d-a6ef-dcd0d0618990","Type":"ContainerStarted","Data":"afc31ce526cf664d9bc4335de653fd9a9349b766fbe06f0d9ad58e2b060772d3"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.131058 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.133093 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.133589 4989 generic.go:334] "Generic (PLEG): container finished" podID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerID="af8fbf4c6bd23d32f4999b8714b7549353870f83d7579e049bbf53f921a862bb" exitCode=0 Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.133675 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerDied","Data":"af8fbf4c6bd23d32f4999b8714b7549353870f83d7579e049bbf53f921a862bb"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.135576 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.136512 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.138304 4989 generic.go:334] "Generic (PLEG): container finished" podID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerID="7a911560c3e865cba1f7875fbb1fc46fe7f621d571b166e361517971d51a95be" exitCode=0 Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.138354 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerDied","Data":"7a911560c3e865cba1f7875fbb1fc46fe7f621d571b166e361517971d51a95be"} Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.145962 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.147354 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.147709 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbwhf\" (UniqueName: \"kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.147786 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.147867 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.148813 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.148924 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.648890713 +0000 UTC m=+160.438916293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.149718 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.161031 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-zlrl6" podStartSLOduration=10.16099063 podStartE2EDuration="10.16099063s" podCreationTimestamp="2025-10-06 08:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:49.152361704 +0000 UTC m=+159.942387294" watchObservedRunningTime="2025-10-06 08:41:49.16099063 +0000 UTC m=+159.951016210" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.170072 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.194835 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbwhf\" (UniqueName: \"kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf\") pod \"redhat-marketplace-5s2bw\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.223903 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.249398 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.249461 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.249524 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.251393 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.751378689 +0000 UTC m=+160.541404269 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.273889 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.274915 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.278196 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.285244 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.350794 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351472 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351551 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351584 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351623 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hx4s\" (UniqueName: \"kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.351793 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.351888 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.851857878 +0000 UTC m=+160.641883478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.372461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.444691 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.450902 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453271 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453358 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hx4s\" (UniqueName: \"kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453414 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.453721 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:49.953704679 +0000 UTC m=+160.743730259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453934 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.453951 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.476628 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.480615 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.477621 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hx4s\" (UniqueName: \"kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s\") pod \"redhat-operators-g8bfp\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.492251 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.554569 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.554685 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.054633246 +0000 UTC m=+160.844658826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.554981 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.555077 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5v7\" (UniqueName: \"kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.555159 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.555207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.555506 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.055490908 +0000 UTC m=+160.845516488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.580809 4989 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.598245 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.630464 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rv274" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.657460 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.657681 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5v7\" (UniqueName: \"kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.657772 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.657795 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.658419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.658488 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.158472962 +0000 UTC m=+160.948498542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.658772 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.689834 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5v7\" (UniqueName: \"kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7\") pod \"redhat-operators-ncs6f\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.712350 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.759550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.759900 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.259889297 +0000 UTC m=+161.049914877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.852237 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.861196 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.861701 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.361682296 +0000 UTC m=+161.151707876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:49 crc kubenswrapper[4989]: I1006 08:41:49.964534 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:49 crc kubenswrapper[4989]: E1006 08:41:49.965104 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.465088526 +0000 UTC m=+161.255114096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p8vh6" (UID: "88a26819-2d38-493d-8792-747304805368") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.031514 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:50 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:50 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:50 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.031564 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.067480 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:50 crc kubenswrapper[4989]: E1006 08:41:50.067817 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-06 08:41:50.56780419 +0000 UTC m=+161.357829760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.081354 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.083340 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.115224 4989 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-06T08:41:49.580831444Z","Handler":null,"Name":""} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.122543 4989 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.122583 4989 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.157233 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b24f0616-9032-4863-a8c4-25fd83ec3b46","Type":"ContainerStarted","Data":"f61ecbfc422e89809204324f603290f214efee96b4e2524166c6a47e45778093"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.157286 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b24f0616-9032-4863-a8c4-25fd83ec3b46","Type":"ContainerStarted","Data":"c799090b22126ce2cb0d7eab9586d43f4f58391d422fde14e289fd072b67c9a6"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.159505 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerStarted","Data":"f2a8cde9736d31b401cb081d61f20f3a98d0634b5732622b06c4e457e3cd85fb"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.163267 4989 generic.go:334] "Generic (PLEG): container finished" podID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerID="a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f" exitCode=0 Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.163344 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerDied","Data":"a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.163481 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerStarted","Data":"b0cc1669ba71a78e7091821ad2b205a4944c5299be6bfdbbc9e167db44f5adb9"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.167062 4989 generic.go:334] "Generic (PLEG): container finished" podID="25ae2464-66b2-4986-beef-382caadb447e" containerID="4d953e5712bd3c81aab1f7118d30d60f3d549a967fbebf044b5f14a3ce7546c3" exitCode=0 Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.167158 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" event={"ID":"25ae2464-66b2-4986-beef-382caadb447e","Type":"ContainerDied","Data":"4d953e5712bd3c81aab1f7118d30d60f3d549a967fbebf044b5f14a3ce7546c3"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.168882 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerStarted","Data":"77043284f006110c2a53a1cf88cc5ae6d985a88295a753cca0c6488a6a65c5a3"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.169157 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.172363 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.172396 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.175009 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.1749961820000001 podStartE2EDuration="1.174996182s" podCreationTimestamp="2025-10-06 08:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:50.17149771 +0000 UTC m=+160.961523290" watchObservedRunningTime="2025-10-06 08:41:50.174996182 +0000 UTC m=+160.965021762" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.192463 4989 generic.go:334] "Generic (PLEG): container finished" podID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerID="a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee" exitCode=0 Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.192702 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerDied","Data":"a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.192735 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerStarted","Data":"128200c39e9872dfd2529145a37f39f937c1fe87eff54d43a3af19ce894b01f5"} Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.244464 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p8vh6\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.270105 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.282568 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.289727 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.553090 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:41:50 crc kubenswrapper[4989]: W1006 08:41:50.560528 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88a26819_2d38_493d_8792_747304805368.slice/crio-4b88b6d4dbde3aa37c2c257e8fe5214f7f2a57cd0b74f9eb48f276f63859cc9f WatchSource:0}: Error finding container 4b88b6d4dbde3aa37c2c257e8fe5214f7f2a57cd0b74f9eb48f276f63859cc9f: Status 404 returned error can't find the container with id 4b88b6d4dbde3aa37c2c257e8fe5214f7f2a57cd0b74f9eb48f276f63859cc9f Oct 06 08:41:50 crc kubenswrapper[4989]: I1006 08:41:50.995624 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.002493 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-87qfj" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.028928 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:51 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:51 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:51 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.028979 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.036948 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-9pfhc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.036996 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9pfhc" podUID="3bd79c86-9633-4dcb-bc5f-800abe66a513" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.037049 4989 patch_prober.go:28] interesting pod/downloads-7954f5f757-9pfhc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.037079 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9pfhc" podUID="3bd79c86-9633-4dcb-bc5f-800abe66a513" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.065132 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.065177 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.066698 4989 patch_prober.go:28] interesting pod/console-f9d7485db-wwr9w container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.066736 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wwr9w" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.210303 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerID="ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84" exitCode=0 Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.210420 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerDied","Data":"ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84"} Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.217499 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" event={"ID":"88a26819-2d38-493d-8792-747304805368","Type":"ContainerStarted","Data":"f90da1cbebdf01715c0b456773a23ccf4928182ddde59d2805d8dcb3a0c24fb3"} Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.217550 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" event={"ID":"88a26819-2d38-493d-8792-747304805368","Type":"ContainerStarted","Data":"4b88b6d4dbde3aa37c2c257e8fe5214f7f2a57cd0b74f9eb48f276f63859cc9f"} Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.217706 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.232076 4989 generic.go:334] "Generic (PLEG): container finished" podID="b24f0616-9032-4863-a8c4-25fd83ec3b46" containerID="f61ecbfc422e89809204324f603290f214efee96b4e2524166c6a47e45778093" exitCode=0 Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.232202 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b24f0616-9032-4863-a8c4-25fd83ec3b46","Type":"ContainerDied","Data":"f61ecbfc422e89809204324f603290f214efee96b4e2524166c6a47e45778093"} Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.256969 4989 generic.go:334] "Generic (PLEG): container finished" podID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerID="c0889d50d9d67470005007e9bb4bdafe51e7ec632b9d9fdcee596fa189478e05" exitCode=0 Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.257163 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerDied","Data":"c0889d50d9d67470005007e9bb4bdafe51e7ec632b9d9fdcee596fa189478e05"} Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.259118 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" podStartSLOduration=137.259104919 podStartE2EDuration="2m17.259104919s" podCreationTimestamp="2025-10-06 08:39:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:41:51.247080756 +0000 UTC m=+162.037106346" watchObservedRunningTime="2025-10-06 08:41:51.259104919 +0000 UTC m=+162.049130499" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.550147 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.611966 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.713961 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume\") pod \"25ae2464-66b2-4986-beef-382caadb447e\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.714112 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume\") pod \"25ae2464-66b2-4986-beef-382caadb447e\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.714175 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmkc9\" (UniqueName: \"kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9\") pod \"25ae2464-66b2-4986-beef-382caadb447e\" (UID: \"25ae2464-66b2-4986-beef-382caadb447e\") " Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.714781 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume" (OuterVolumeSpecName: "config-volume") pod "25ae2464-66b2-4986-beef-382caadb447e" (UID: "25ae2464-66b2-4986-beef-382caadb447e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.717908 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25ae2464-66b2-4986-beef-382caadb447e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.726452 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "25ae2464-66b2-4986-beef-382caadb447e" (UID: "25ae2464-66b2-4986-beef-382caadb447e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.734544 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9" (OuterVolumeSpecName: "kube-api-access-dmkc9") pod "25ae2464-66b2-4986-beef-382caadb447e" (UID: "25ae2464-66b2-4986-beef-382caadb447e"). InnerVolumeSpecName "kube-api-access-dmkc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.819705 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmkc9\" (UniqueName: \"kubernetes.io/projected/25ae2464-66b2-4986-beef-382caadb447e-kube-api-access-dmkc9\") on node \"crc\" DevicePath \"\"" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.819750 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25ae2464-66b2-4986-beef-382caadb447e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 08:41:51 crc kubenswrapper[4989]: I1006 08:41:51.956258 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.026326 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.029461 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:52 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:52 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:52 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.029507 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.275278 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" event={"ID":"25ae2464-66b2-4986-beef-382caadb447e","Type":"ContainerDied","Data":"e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680"} Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.275326 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e74ec905985744d0cb76b33c1dc8abae7c29425d3ec6ec81a164902dafbe6680" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.275474 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.707468 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.838470 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir\") pod \"b24f0616-9032-4863-a8c4-25fd83ec3b46\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.838614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access\") pod \"b24f0616-9032-4863-a8c4-25fd83ec3b46\" (UID: \"b24f0616-9032-4863-a8c4-25fd83ec3b46\") " Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.839328 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b24f0616-9032-4863-a8c4-25fd83ec3b46" (UID: "b24f0616-9032-4863-a8c4-25fd83ec3b46"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.847404 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b24f0616-9032-4863-a8c4-25fd83ec3b46" (UID: "b24f0616-9032-4863-a8c4-25fd83ec3b46"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.940267 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b24f0616-9032-4863-a8c4-25fd83ec3b46-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:41:52 crc kubenswrapper[4989]: I1006 08:41:52.940295 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b24f0616-9032-4863-a8c4-25fd83ec3b46-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.028426 4989 patch_prober.go:28] interesting pod/router-default-5444994796-c829n container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 06 08:41:53 crc kubenswrapper[4989]: [-]has-synced failed: reason withheld Oct 06 08:41:53 crc kubenswrapper[4989]: [+]process-running ok Oct 06 08:41:53 crc kubenswrapper[4989]: healthz check failed Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.028494 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-c829n" podUID="bd7241b3-d77b-458e-b4bf-77fbd9f28596" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.253669 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.291214 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b24f0616-9032-4863-a8c4-25fd83ec3b46","Type":"ContainerDied","Data":"c799090b22126ce2cb0d7eab9586d43f4f58391d422fde14e289fd072b67c9a6"} Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.291259 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c799090b22126ce2cb0d7eab9586d43f4f58391d422fde14e289fd072b67c9a6" Oct 06 08:41:53 crc kubenswrapper[4989]: I1006 08:41:53.291276 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.028594 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.031014 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-c829n" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.066857 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x9rxz" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.397320 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 08:41:54 crc kubenswrapper[4989]: E1006 08:41:54.397536 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24f0616-9032-4863-a8c4-25fd83ec3b46" containerName="pruner" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.397547 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24f0616-9032-4863-a8c4-25fd83ec3b46" containerName="pruner" Oct 06 08:41:54 crc kubenswrapper[4989]: E1006 08:41:54.397558 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ae2464-66b2-4986-beef-382caadb447e" containerName="collect-profiles" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.397564 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ae2464-66b2-4986-beef-382caadb447e" containerName="collect-profiles" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.397704 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24f0616-9032-4863-a8c4-25fd83ec3b46" containerName="pruner" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.397717 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ae2464-66b2-4986-beef-382caadb447e" containerName="collect-profiles" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.398142 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.402312 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.402501 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.403934 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.466826 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.466953 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.568071 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.568469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.568539 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.600383 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:54 crc kubenswrapper[4989]: I1006 08:41:54.721911 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:41:56 crc kubenswrapper[4989]: I1006 08:41:56.900168 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:56 crc kubenswrapper[4989]: I1006 08:41:56.906532 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c9cca673-1746-4fd9-bc67-330c4d06522e-metrics-certs\") pod \"network-metrics-daemon-8ndcb\" (UID: \"c9cca673-1746-4fd9-bc67-330c4d06522e\") " pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:41:57 crc kubenswrapper[4989]: I1006 08:41:57.059250 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8ndcb" Oct 06 08:42:01 crc kubenswrapper[4989]: I1006 08:42:01.042413 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9pfhc" Oct 06 08:42:01 crc kubenswrapper[4989]: I1006 08:42:01.084437 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:42:01 crc kubenswrapper[4989]: I1006 08:42:01.089887 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:42:03 crc kubenswrapper[4989]: I1006 08:42:03.935419 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:42:03 crc kubenswrapper[4989]: I1006 08:42:03.935774 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:42:07 crc kubenswrapper[4989]: I1006 08:42:07.325866 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 06 08:42:08 crc kubenswrapper[4989]: I1006 08:42:08.023563 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 06 08:42:10 crc kubenswrapper[4989]: I1006 08:42:10.296064 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:42:13 crc kubenswrapper[4989]: E1006 08:42:13.822631 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 06 08:42:13 crc kubenswrapper[4989]: E1006 08:42:13.823292 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tl5v7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-ncs6f_openshift-marketplace(7f49b107-bb1a-49c1-ab50-fad36694e25b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:13 crc kubenswrapper[4989]: E1006 08:42:13.824646 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-ncs6f" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" Oct 06 08:42:16 crc kubenswrapper[4989]: W1006 08:42:16.924897 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc79f5da6_1ccb_4e02_83ca_772969378d20.slice/crio-6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078 WatchSource:0}: Error finding container 6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078: Status 404 returned error can't find the container with id 6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078 Oct 06 08:42:17 crc kubenswrapper[4989]: E1006 08:42:17.020428 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 06 08:42:17 crc kubenswrapper[4989]: E1006 08:42:17.020940 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wb5sh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rqnb9_openshift-marketplace(9e7970d8-5509-4ec1-a9e0-549db8358313): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:17 crc kubenswrapper[4989]: E1006 08:42:17.022268 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rqnb9" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" Oct 06 08:42:17 crc kubenswrapper[4989]: I1006 08:42:17.431940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c79f5da6-1ccb-4e02-83ca-772969378d20","Type":"ContainerStarted","Data":"6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078"} Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.113601 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-ncs6f" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.113721 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rqnb9" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.198749 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.198953 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xdwd4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7xfhm_openshift-marketplace(e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.200126 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7xfhm" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.201184 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.201324 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pp8wh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9rl5s_openshift-marketplace(fa58da1b-2409-49a0-a2f3-55e83d584a91): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.202990 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9rl5s" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.209808 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.209947 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xl7g2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tnvzq_openshift-marketplace(e90ef368-0b6b-4c6a-b96f-7530016bc537): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.211109 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tnvzq" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.228382 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.228501 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4hx4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-g8bfp_openshift-marketplace(092ad7d4-11e3-4399-a9ae-65f548ba3f85): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:18 crc kubenswrapper[4989]: E1006 08:42:18.229680 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-g8bfp" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" Oct 06 08:42:20 crc kubenswrapper[4989]: E1006 08:42:20.797571 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-g8bfp" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" Oct 06 08:42:20 crc kubenswrapper[4989]: E1006 08:42:20.797733 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7xfhm" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" Oct 06 08:42:20 crc kubenswrapper[4989]: E1006 08:42:20.797832 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tnvzq" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" Oct 06 08:42:20 crc kubenswrapper[4989]: E1006 08:42:20.797872 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9rl5s" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" Oct 06 08:42:20 crc kubenswrapper[4989]: I1006 08:42:20.961912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8ndcb"] Oct 06 08:42:21 crc kubenswrapper[4989]: I1006 08:42:21.637587 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2h8tc" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.726513 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.726701 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbwhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5s2bw_openshift-marketplace(c1fd2533-7e30-4022-b76b-02a354cdb2d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.727861 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5s2bw" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.788326 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.788549 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f8xzl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4q4mm_openshift-marketplace(92ea6a88-1fac-4940-96e8-ec0a33d9951f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 06 08:42:21 crc kubenswrapper[4989]: E1006 08:42:21.789713 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4q4mm" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.459196 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" event={"ID":"c9cca673-1746-4fd9-bc67-330c4d06522e","Type":"ContainerStarted","Data":"76e2d9bd30ba610f04362a4c0a0676f44b706a8601dae0664e46bb79d8221000"} Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.459717 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" event={"ID":"c9cca673-1746-4fd9-bc67-330c4d06522e","Type":"ContainerStarted","Data":"7c9fd19d4c987deadeae4f6624c040cfc7cb956faf54c75174b7dd68c2ce3b6f"} Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.459747 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8ndcb" event={"ID":"c9cca673-1746-4fd9-bc67-330c4d06522e","Type":"ContainerStarted","Data":"5c9ce1e52e3debc5aaef57e2b5a20e8b71654426a1640929ab5faf629c5a2bdf"} Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.462107 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c79f5da6-1ccb-4e02-83ca-772969378d20","Type":"ContainerStarted","Data":"466da47de3f007f1a7da436a82b61f1a79ffdf12f97052000cfd94d28377cea5"} Oct 06 08:42:22 crc kubenswrapper[4989]: E1006 08:42:22.466379 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5s2bw" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" Oct 06 08:42:22 crc kubenswrapper[4989]: E1006 08:42:22.466416 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4q4mm" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.477791 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8ndcb" podStartSLOduration=169.477770493 podStartE2EDuration="2m49.477770493s" podCreationTimestamp="2025-10-06 08:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:42:22.47661708 +0000 UTC m=+193.266642710" watchObservedRunningTime="2025-10-06 08:42:22.477770493 +0000 UTC m=+193.267796103" Oct 06 08:42:22 crc kubenswrapper[4989]: I1006 08:42:22.520068 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=28.520040118 podStartE2EDuration="28.520040118s" podCreationTimestamp="2025-10-06 08:41:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:42:22.51347719 +0000 UTC m=+193.303502780" watchObservedRunningTime="2025-10-06 08:42:22.520040118 +0000 UTC m=+193.310065738" Oct 06 08:42:23 crc kubenswrapper[4989]: I1006 08:42:23.470496 4989 generic.go:334] "Generic (PLEG): container finished" podID="c79f5da6-1ccb-4e02-83ca-772969378d20" containerID="466da47de3f007f1a7da436a82b61f1a79ffdf12f97052000cfd94d28377cea5" exitCode=0 Oct 06 08:42:23 crc kubenswrapper[4989]: I1006 08:42:23.470594 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c79f5da6-1ccb-4e02-83ca-772969378d20","Type":"ContainerDied","Data":"466da47de3f007f1a7da436a82b61f1a79ffdf12f97052000cfd94d28377cea5"} Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.737731 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.778139 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access\") pod \"c79f5da6-1ccb-4e02-83ca-772969378d20\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.778199 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir\") pod \"c79f5da6-1ccb-4e02-83ca-772969378d20\" (UID: \"c79f5da6-1ccb-4e02-83ca-772969378d20\") " Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.778477 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c79f5da6-1ccb-4e02-83ca-772969378d20" (UID: "c79f5da6-1ccb-4e02-83ca-772969378d20"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.793279 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c79f5da6-1ccb-4e02-83ca-772969378d20" (UID: "c79f5da6-1ccb-4e02-83ca-772969378d20"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.879219 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c79f5da6-1ccb-4e02-83ca-772969378d20-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:24 crc kubenswrapper[4989]: I1006 08:42:24.879249 4989 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c79f5da6-1ccb-4e02-83ca-772969378d20-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:25 crc kubenswrapper[4989]: I1006 08:42:25.482547 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c79f5da6-1ccb-4e02-83ca-772969378d20","Type":"ContainerDied","Data":"6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078"} Oct 06 08:42:25 crc kubenswrapper[4989]: I1006 08:42:25.482594 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6532c16e6447a3408f6b8418207bdd5a99a65531c6e74e365a3cbb189b8ef078" Oct 06 08:42:25 crc kubenswrapper[4989]: I1006 08:42:25.482699 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 06 08:42:31 crc kubenswrapper[4989]: I1006 08:42:31.521335 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerStarted","Data":"9bb28c10e88fc5da8bcdd5ae5be9631c448d8a444ccee4d22afbbd7630fb5007"} Oct 06 08:42:31 crc kubenswrapper[4989]: I1006 08:42:31.522685 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerStarted","Data":"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2"} Oct 06 08:42:32 crc kubenswrapper[4989]: I1006 08:42:32.531585 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerID="9bb28c10e88fc5da8bcdd5ae5be9631c448d8a444ccee4d22afbbd7630fb5007" exitCode=0 Oct 06 08:42:32 crc kubenswrapper[4989]: I1006 08:42:32.531839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerDied","Data":"9bb28c10e88fc5da8bcdd5ae5be9631c448d8a444ccee4d22afbbd7630fb5007"} Oct 06 08:42:32 crc kubenswrapper[4989]: I1006 08:42:32.536712 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerID="51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2" exitCode=0 Oct 06 08:42:32 crc kubenswrapper[4989]: I1006 08:42:32.536778 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerDied","Data":"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2"} Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.545569 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerID="256dc12dc3f879ba5e3fbc593e86b612e5bb3211ad01af3a19ab77f456cd14cd" exitCode=0 Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.545697 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerDied","Data":"256dc12dc3f879ba5e3fbc593e86b612e5bb3211ad01af3a19ab77f456cd14cd"} Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.551275 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerStarted","Data":"b977114263be54860df340e3891de10f05e2b5318b0f53bab1b59778137eae14"} Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.554555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerStarted","Data":"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726"} Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.583371 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ncs6f" podStartSLOduration=2.760054092 podStartE2EDuration="44.58335256s" podCreationTimestamp="2025-10-06 08:41:49 +0000 UTC" firstStartedPulling="2025-10-06 08:41:51.212032644 +0000 UTC m=+162.002058224" lastFinishedPulling="2025-10-06 08:42:33.035331112 +0000 UTC m=+203.825356692" observedRunningTime="2025-10-06 08:42:33.580612648 +0000 UTC m=+204.370638228" watchObservedRunningTime="2025-10-06 08:42:33.58335256 +0000 UTC m=+204.373378140" Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.599405 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqnb9" podStartSLOduration=3.690293271 podStartE2EDuration="47.599384149s" podCreationTimestamp="2025-10-06 08:41:46 +0000 UTC" firstStartedPulling="2025-10-06 08:41:49.105169244 +0000 UTC m=+159.895194814" lastFinishedPulling="2025-10-06 08:42:33.014260112 +0000 UTC m=+203.804285692" observedRunningTime="2025-10-06 08:42:33.596917015 +0000 UTC m=+204.386942595" watchObservedRunningTime="2025-10-06 08:42:33.599384149 +0000 UTC m=+204.389409739" Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.935889 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:42:33 crc kubenswrapper[4989]: I1006 08:42:33.935948 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:42:34 crc kubenswrapper[4989]: I1006 08:42:34.563139 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerStarted","Data":"2c2c34770979a8f74c07197273733e4ef2e902e41a2edb5f5bdd72d36f0473c0"} Oct 06 08:42:34 crc kubenswrapper[4989]: I1006 08:42:34.588777 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9rl5s" podStartSLOduration=3.561970329 podStartE2EDuration="48.58875381s" podCreationTimestamp="2025-10-06 08:41:46 +0000 UTC" firstStartedPulling="2025-10-06 08:41:49.098931219 +0000 UTC m=+159.888956799" lastFinishedPulling="2025-10-06 08:42:34.1257147 +0000 UTC m=+204.915740280" observedRunningTime="2025-10-06 08:42:34.586305276 +0000 UTC m=+205.376330866" watchObservedRunningTime="2025-10-06 08:42:34.58875381 +0000 UTC m=+205.378779390" Oct 06 08:42:35 crc kubenswrapper[4989]: I1006 08:42:35.574192 4989 generic.go:334] "Generic (PLEG): container finished" podID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerID="91ff25207f07f96ec836fae08a23e60f7c522bb36f807193a86b1ac58b3d8682" exitCode=0 Oct 06 08:42:35 crc kubenswrapper[4989]: I1006 08:42:35.574374 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerDied","Data":"91ff25207f07f96ec836fae08a23e60f7c522bb36f807193a86b1ac58b3d8682"} Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.581525 4989 generic.go:334] "Generic (PLEG): container finished" podID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerID="2df10fe20f69a9a91e7d5091a3aa45e18def066b5bb4e7f12f6dd8830dbe1a58" exitCode=0 Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.581755 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerDied","Data":"2df10fe20f69a9a91e7d5091a3aa45e18def066b5bb4e7f12f6dd8830dbe1a58"} Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.586818 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerStarted","Data":"c242062bf0407096034796a0010bd22528b184749c4859e0b9792ee6a33f6bf5"} Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.589049 4989 generic.go:334] "Generic (PLEG): container finished" podID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerID="38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633" exitCode=0 Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.589075 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerDied","Data":"38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633"} Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.633883 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7xfhm" podStartSLOduration=3.662097808 podStartE2EDuration="50.633868382s" podCreationTimestamp="2025-10-06 08:41:46 +0000 UTC" firstStartedPulling="2025-10-06 08:41:49.139481539 +0000 UTC m=+159.929507119" lastFinishedPulling="2025-10-06 08:42:36.111252113 +0000 UTC m=+206.901277693" observedRunningTime="2025-10-06 08:42:36.631598915 +0000 UTC m=+207.421624495" watchObservedRunningTime="2025-10-06 08:42:36.633868382 +0000 UTC m=+207.423893962" Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.949969 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:36 crc kubenswrapper[4989]: I1006 08:42:36.950002 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.044224 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.044776 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.090019 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.093227 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.594984 4989 generic.go:334] "Generic (PLEG): container finished" podID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerID="d703d13a9adb8cc784c416b62c5901c76d0c805e2fa0a339689d98d8c40d5f8b" exitCode=0 Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.595051 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerDied","Data":"d703d13a9adb8cc784c416b62c5901c76d0c805e2fa0a339689d98d8c40d5f8b"} Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.596444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerStarted","Data":"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5"} Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.599280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerStarted","Data":"1c24ced7cecfdb737a3624fd3c2d2854604b77c02df3c656884d33efddf2bfb4"} Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.643468 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4q4mm" podStartSLOduration=2.860544382 podStartE2EDuration="49.643450337s" podCreationTimestamp="2025-10-06 08:41:48 +0000 UTC" firstStartedPulling="2025-10-06 08:41:50.197103516 +0000 UTC m=+160.987129096" lastFinishedPulling="2025-10-06 08:42:36.980009471 +0000 UTC m=+207.770035051" observedRunningTime="2025-10-06 08:42:37.642017114 +0000 UTC m=+208.432042694" watchObservedRunningTime="2025-10-06 08:42:37.643450337 +0000 UTC m=+208.433475917" Oct 06 08:42:37 crc kubenswrapper[4989]: I1006 08:42:37.662547 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8bfp" podStartSLOduration=2.735987249 podStartE2EDuration="48.662534027s" podCreationTimestamp="2025-10-06 08:41:49 +0000 UTC" firstStartedPulling="2025-10-06 08:41:51.261237239 +0000 UTC m=+162.051262819" lastFinishedPulling="2025-10-06 08:42:37.187784017 +0000 UTC m=+207.977809597" observedRunningTime="2025-10-06 08:42:37.660582839 +0000 UTC m=+208.450608419" watchObservedRunningTime="2025-10-06 08:42:37.662534027 +0000 UTC m=+208.452559607" Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.607582 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerStarted","Data":"e5025a816f27f9acce03351c576f8e7e2baca27fffa094f6d522df42603bb744"} Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.609743 4989 generic.go:334] "Generic (PLEG): container finished" podID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerID="9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321" exitCode=0 Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.609799 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerDied","Data":"9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321"} Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.635110 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tnvzq" podStartSLOduration=3.6094001799999997 podStartE2EDuration="52.635092326s" podCreationTimestamp="2025-10-06 08:41:46 +0000 UTC" firstStartedPulling="2025-10-06 08:41:49.136287678 +0000 UTC m=+159.926313258" lastFinishedPulling="2025-10-06 08:42:38.161979814 +0000 UTC m=+208.952005404" observedRunningTime="2025-10-06 08:42:38.627152878 +0000 UTC m=+209.417178478" watchObservedRunningTime="2025-10-06 08:42:38.635092326 +0000 UTC m=+209.425117916" Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.845803 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.846072 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:42:38 crc kubenswrapper[4989]: I1006 08:42:38.890685 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.599004 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.599349 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.617562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerStarted","Data":"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e"} Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.641113 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5s2bw" podStartSLOduration=2.586190627 podStartE2EDuration="51.641092192s" podCreationTimestamp="2025-10-06 08:41:48 +0000 UTC" firstStartedPulling="2025-10-06 08:41:50.165156551 +0000 UTC m=+160.955182131" lastFinishedPulling="2025-10-06 08:42:39.220058116 +0000 UTC m=+210.010083696" observedRunningTime="2025-10-06 08:42:39.636750152 +0000 UTC m=+210.426775732" watchObservedRunningTime="2025-10-06 08:42:39.641092192 +0000 UTC m=+210.431117772" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.853138 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.853452 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:39 crc kubenswrapper[4989]: I1006 08:42:39.889862 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:40 crc kubenswrapper[4989]: I1006 08:42:40.633960 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8bfp" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="registry-server" probeResult="failure" output=< Oct 06 08:42:40 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 08:42:40 crc kubenswrapper[4989]: > Oct 06 08:42:40 crc kubenswrapper[4989]: I1006 08:42:40.682782 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:43 crc kubenswrapper[4989]: I1006 08:42:43.166241 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:42:43 crc kubenswrapper[4989]: I1006 08:42:43.166502 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ncs6f" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="registry-server" containerID="cri-o://e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726" gracePeriod=2 Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.509271 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.531995 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities\") pod \"7f49b107-bb1a-49c1-ab50-fad36694e25b\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.532067 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content\") pod \"7f49b107-bb1a-49c1-ab50-fad36694e25b\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.532148 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl5v7\" (UniqueName: \"kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7\") pod \"7f49b107-bb1a-49c1-ab50-fad36694e25b\" (UID: \"7f49b107-bb1a-49c1-ab50-fad36694e25b\") " Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.533407 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities" (OuterVolumeSpecName: "utilities") pod "7f49b107-bb1a-49c1-ab50-fad36694e25b" (UID: "7f49b107-bb1a-49c1-ab50-fad36694e25b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.537552 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7" (OuterVolumeSpecName: "kube-api-access-tl5v7") pod "7f49b107-bb1a-49c1-ab50-fad36694e25b" (UID: "7f49b107-bb1a-49c1-ab50-fad36694e25b"). InnerVolumeSpecName "kube-api-access-tl5v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.612385 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f49b107-bb1a-49c1-ab50-fad36694e25b" (UID: "7f49b107-bb1a-49c1-ab50-fad36694e25b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.633130 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.633172 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f49b107-bb1a-49c1-ab50-fad36694e25b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.633189 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl5v7\" (UniqueName: \"kubernetes.io/projected/7f49b107-bb1a-49c1-ab50-fad36694e25b-kube-api-access-tl5v7\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.644384 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerID="e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726" exitCode=0 Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.644415 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ncs6f" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.644437 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerDied","Data":"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726"} Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.644497 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ncs6f" event={"ID":"7f49b107-bb1a-49c1-ab50-fad36694e25b","Type":"ContainerDied","Data":"77043284f006110c2a53a1cf88cc5ae6d985a88295a753cca0c6488a6a65c5a3"} Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.644519 4989 scope.go:117] "RemoveContainer" containerID="e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.660208 4989 scope.go:117] "RemoveContainer" containerID="51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.673543 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.677324 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ncs6f"] Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.695700 4989 scope.go:117] "RemoveContainer" containerID="ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.711751 4989 scope.go:117] "RemoveContainer" containerID="e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726" Oct 06 08:42:44 crc kubenswrapper[4989]: E1006 08:42:44.712263 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726\": container with ID starting with e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726 not found: ID does not exist" containerID="e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.712318 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726"} err="failed to get container status \"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726\": rpc error: code = NotFound desc = could not find container \"e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726\": container with ID starting with e817bba541fd058810e76e19253ae11c046773791ebaf0021240538eb3969726 not found: ID does not exist" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.712401 4989 scope.go:117] "RemoveContainer" containerID="51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2" Oct 06 08:42:44 crc kubenswrapper[4989]: E1006 08:42:44.712878 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2\": container with ID starting with 51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2 not found: ID does not exist" containerID="51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.712928 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2"} err="failed to get container status \"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2\": rpc error: code = NotFound desc = could not find container \"51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2\": container with ID starting with 51ed13a95c6c864decdb964349244d8ce73544bff434305bf5af2bf41f13b0a2 not found: ID does not exist" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.712968 4989 scope.go:117] "RemoveContainer" containerID="ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84" Oct 06 08:42:44 crc kubenswrapper[4989]: E1006 08:42:44.713288 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84\": container with ID starting with ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84 not found: ID does not exist" containerID="ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84" Oct 06 08:42:44 crc kubenswrapper[4989]: I1006 08:42:44.713323 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84"} err="failed to get container status \"ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84\": rpc error: code = NotFound desc = could not find container \"ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84\": container with ID starting with ede6078c2a8508c3a0c75ce7adb8cd182f319f1f767cc3bc269075c1ffe08c84 not found: ID does not exist" Oct 06 08:42:45 crc kubenswrapper[4989]: I1006 08:42:45.941471 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" path="/var/lib/kubelet/pods/7f49b107-bb1a-49c1-ab50-fad36694e25b/volumes" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.479460 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.479501 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.515646 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.619175 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.619242 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.687087 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.699179 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.733687 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:42:46 crc kubenswrapper[4989]: I1006 08:42:46.989569 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:47 crc kubenswrapper[4989]: I1006 08:42:47.090360 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.368093 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.368433 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9rl5s" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="registry-server" containerID="cri-o://2c2c34770979a8f74c07197273733e4ef2e902e41a2edb5f5bdd72d36f0473c0" gracePeriod=2 Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.670582 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerID="2c2c34770979a8f74c07197273733e4ef2e902e41a2edb5f5bdd72d36f0473c0" exitCode=0 Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.670625 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerDied","Data":"2c2c34770979a8f74c07197273733e4ef2e902e41a2edb5f5bdd72d36f0473c0"} Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.886327 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.965097 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:42:48 crc kubenswrapper[4989]: I1006 08:42:48.965323 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqnb9" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="registry-server" containerID="cri-o://b977114263be54860df340e3891de10f05e2b5318b0f53bab1b59778137eae14" gracePeriod=2 Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.224869 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.224925 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.263731 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.428691 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.503637 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp8wh\" (UniqueName: \"kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh\") pod \"fa58da1b-2409-49a0-a2f3-55e83d584a91\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.503732 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content\") pod \"fa58da1b-2409-49a0-a2f3-55e83d584a91\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.503763 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities\") pod \"fa58da1b-2409-49a0-a2f3-55e83d584a91\" (UID: \"fa58da1b-2409-49a0-a2f3-55e83d584a91\") " Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.504668 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities" (OuterVolumeSpecName: "utilities") pod "fa58da1b-2409-49a0-a2f3-55e83d584a91" (UID: "fa58da1b-2409-49a0-a2f3-55e83d584a91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.510377 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh" (OuterVolumeSpecName: "kube-api-access-pp8wh") pod "fa58da1b-2409-49a0-a2f3-55e83d584a91" (UID: "fa58da1b-2409-49a0-a2f3-55e83d584a91"). InnerVolumeSpecName "kube-api-access-pp8wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.547870 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa58da1b-2409-49a0-a2f3-55e83d584a91" (UID: "fa58da1b-2409-49a0-a2f3-55e83d584a91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.604808 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp8wh\" (UniqueName: \"kubernetes.io/projected/fa58da1b-2409-49a0-a2f3-55e83d584a91-kube-api-access-pp8wh\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.605324 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.605413 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58da1b-2409-49a0-a2f3-55e83d584a91-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.641855 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.683377 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rl5s" event={"ID":"fa58da1b-2409-49a0-a2f3-55e83d584a91","Type":"ContainerDied","Data":"99894760e93a8d92cd98ff9b27cece2618eeee3116230f6e4e601a1183b3a646"} Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.683392 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rl5s" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.683485 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.683523 4989 scope.go:117] "RemoveContainer" containerID="2c2c34770979a8f74c07197273733e4ef2e902e41a2edb5f5bdd72d36f0473c0" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.686759 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerID="b977114263be54860df340e3891de10f05e2b5318b0f53bab1b59778137eae14" exitCode=0 Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.686805 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerDied","Data":"b977114263be54860df340e3891de10f05e2b5318b0f53bab1b59778137eae14"} Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.714743 4989 scope.go:117] "RemoveContainer" containerID="256dc12dc3f879ba5e3fbc593e86b612e5bb3211ad01af3a19ab77f456cd14cd" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.728342 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.732919 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.738127 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9rl5s"] Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.738476 4989 scope.go:117] "RemoveContainer" containerID="973ad1862e4fbcc7bffb26582189156a7baa09337bd3ab034e2f5f420c94edfd" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.941136 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" path="/var/lib/kubelet/pods/fa58da1b-2409-49a0-a2f3-55e83d584a91/volumes" Oct 06 08:42:49 crc kubenswrapper[4989]: I1006 08:42:49.958330 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.012532 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities\") pod \"9e7970d8-5509-4ec1-a9e0-549db8358313\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.012597 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb5sh\" (UniqueName: \"kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh\") pod \"9e7970d8-5509-4ec1-a9e0-549db8358313\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.012766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content\") pod \"9e7970d8-5509-4ec1-a9e0-549db8358313\" (UID: \"9e7970d8-5509-4ec1-a9e0-549db8358313\") " Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.013561 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities" (OuterVolumeSpecName: "utilities") pod "9e7970d8-5509-4ec1-a9e0-549db8358313" (UID: "9e7970d8-5509-4ec1-a9e0-549db8358313"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.013942 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.016729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh" (OuterVolumeSpecName: "kube-api-access-wb5sh") pod "9e7970d8-5509-4ec1-a9e0-549db8358313" (UID: "9e7970d8-5509-4ec1-a9e0-549db8358313"). InnerVolumeSpecName "kube-api-access-wb5sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.092032 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e7970d8-5509-4ec1-a9e0-549db8358313" (UID: "9e7970d8-5509-4ec1-a9e0-549db8358313"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.114793 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e7970d8-5509-4ec1-a9e0-549db8358313-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.114833 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb5sh\" (UniqueName: \"kubernetes.io/projected/9e7970d8-5509-4ec1-a9e0-549db8358313-kube-api-access-wb5sh\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.694591 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqnb9" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.694589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqnb9" event={"ID":"9e7970d8-5509-4ec1-a9e0-549db8358313","Type":"ContainerDied","Data":"a7037daea8a158c5a7b7e1c4d7e9ee92a4caaf5c3191ddc39c7d19059a517c0c"} Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.694726 4989 scope.go:117] "RemoveContainer" containerID="b977114263be54860df340e3891de10f05e2b5318b0f53bab1b59778137eae14" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.711605 4989 scope.go:117] "RemoveContainer" containerID="9bb28c10e88fc5da8bcdd5ae5be9631c448d8a444ccee4d22afbbd7630fb5007" Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.722146 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.727031 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqnb9"] Oct 06 08:42:50 crc kubenswrapper[4989]: I1006 08:42:50.756045 4989 scope.go:117] "RemoveContainer" containerID="091dd380f97b2731ca1b118960f0f724454c122308c32eb9f618abd2b9274e2d" Oct 06 08:42:51 crc kubenswrapper[4989]: I1006 08:42:51.367990 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:42:51 crc kubenswrapper[4989]: I1006 08:42:51.701217 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5s2bw" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="registry-server" containerID="cri-o://646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e" gracePeriod=2 Oct 06 08:42:51 crc kubenswrapper[4989]: E1006 08:42:51.734495 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1fd2533_7e30_4022_b76b_02a354cdb2d0.slice/crio-646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e.scope\": RecentStats: unable to find data in memory cache]" Oct 06 08:42:51 crc kubenswrapper[4989]: I1006 08:42:51.942877 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" path="/var/lib/kubelet/pods/9e7970d8-5509-4ec1-a9e0-549db8358313/volumes" Oct 06 08:42:51 crc kubenswrapper[4989]: I1006 08:42:51.997754 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.138438 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content\") pod \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.138518 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities\") pod \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.138617 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbwhf\" (UniqueName: \"kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf\") pod \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\" (UID: \"c1fd2533-7e30-4022-b76b-02a354cdb2d0\") " Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.140054 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities" (OuterVolumeSpecName: "utilities") pod "c1fd2533-7e30-4022-b76b-02a354cdb2d0" (UID: "c1fd2533-7e30-4022-b76b-02a354cdb2d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.143893 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf" (OuterVolumeSpecName: "kube-api-access-xbwhf") pod "c1fd2533-7e30-4022-b76b-02a354cdb2d0" (UID: "c1fd2533-7e30-4022-b76b-02a354cdb2d0"). InnerVolumeSpecName "kube-api-access-xbwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.150745 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1fd2533-7e30-4022-b76b-02a354cdb2d0" (UID: "c1fd2533-7e30-4022-b76b-02a354cdb2d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.240474 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.240731 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1fd2533-7e30-4022-b76b-02a354cdb2d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.240742 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbwhf\" (UniqueName: \"kubernetes.io/projected/c1fd2533-7e30-4022-b76b-02a354cdb2d0-kube-api-access-xbwhf\") on node \"crc\" DevicePath \"\"" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.709551 4989 generic.go:334] "Generic (PLEG): container finished" podID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerID="646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e" exitCode=0 Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.709608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerDied","Data":"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e"} Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.710691 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5s2bw" event={"ID":"c1fd2533-7e30-4022-b76b-02a354cdb2d0","Type":"ContainerDied","Data":"b0cc1669ba71a78e7091821ad2b205a4944c5299be6bfdbbc9e167db44f5adb9"} Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.709697 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5s2bw" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.710736 4989 scope.go:117] "RemoveContainer" containerID="646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.731772 4989 scope.go:117] "RemoveContainer" containerID="9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.746156 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.750543 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5s2bw"] Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.759817 4989 scope.go:117] "RemoveContainer" containerID="a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.780838 4989 scope.go:117] "RemoveContainer" containerID="646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e" Oct 06 08:42:52 crc kubenswrapper[4989]: E1006 08:42:52.781329 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e\": container with ID starting with 646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e not found: ID does not exist" containerID="646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.781374 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e"} err="failed to get container status \"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e\": rpc error: code = NotFound desc = could not find container \"646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e\": container with ID starting with 646a0418fa51318ed737b39e046487fb55193c241e084a7c67788ebcefcff65e not found: ID does not exist" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.781402 4989 scope.go:117] "RemoveContainer" containerID="9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321" Oct 06 08:42:52 crc kubenswrapper[4989]: E1006 08:42:52.781845 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321\": container with ID starting with 9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321 not found: ID does not exist" containerID="9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.781879 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321"} err="failed to get container status \"9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321\": rpc error: code = NotFound desc = could not find container \"9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321\": container with ID starting with 9c5ca5a3ecd83f9997c9f44de1b69790327ae18e20ebd24671e073915d692321 not found: ID does not exist" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.781901 4989 scope.go:117] "RemoveContainer" containerID="a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f" Oct 06 08:42:52 crc kubenswrapper[4989]: E1006 08:42:52.782191 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f\": container with ID starting with a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f not found: ID does not exist" containerID="a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f" Oct 06 08:42:52 crc kubenswrapper[4989]: I1006 08:42:52.782219 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f"} err="failed to get container status \"a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f\": rpc error: code = NotFound desc = could not find container \"a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f\": container with ID starting with a350144b7c84a04acc14f0c536285a61430d2954355f330854a5273f0b8b808f not found: ID does not exist" Oct 06 08:42:53 crc kubenswrapper[4989]: I1006 08:42:53.942008 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" path="/var/lib/kubelet/pods/c1fd2533-7e30-4022-b76b-02a354cdb2d0/volumes" Oct 06 08:43:00 crc kubenswrapper[4989]: I1006 08:43:00.777017 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:43:03 crc kubenswrapper[4989]: I1006 08:43:03.935164 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:43:03 crc kubenswrapper[4989]: I1006 08:43:03.935539 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:43:03 crc kubenswrapper[4989]: I1006 08:43:03.942984 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:43:03 crc kubenswrapper[4989]: I1006 08:43:03.943581 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 08:43:03 crc kubenswrapper[4989]: I1006 08:43:03.943641 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68" gracePeriod=600 Oct 06 08:43:04 crc kubenswrapper[4989]: I1006 08:43:04.775639 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68" exitCode=0 Oct 06 08:43:04 crc kubenswrapper[4989]: I1006 08:43:04.775719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68"} Oct 06 08:43:04 crc kubenswrapper[4989]: I1006 08:43:04.776199 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97"} Oct 06 08:43:25 crc kubenswrapper[4989]: I1006 08:43:25.804871 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" podUID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" containerName="oauth-openshift" containerID="cri-o://10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b" gracePeriod=15 Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.132122 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172105 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4tpc\" (UniqueName: \"kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172174 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172195 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172216 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172235 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172263 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172296 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172339 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172365 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172416 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172435 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172454 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172473 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.172495 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection\") pod \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\" (UID: \"14cad03b-9a44-4f7f-a3a1-ae8b970fd993\") " Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.173213 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.173420 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.173453 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.173475 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.173926 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180281 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-cr2rf"] Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180486 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180497 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180507 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180514 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180531 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" containerName="oauth-openshift" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180536 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" containerName="oauth-openshift" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180547 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180552 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180561 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180567 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180574 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180581 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180587 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180592 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180599 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180604 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180612 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180618 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180625 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180630 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180636 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180694 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="extract-content" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180705 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180711 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180719 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180725 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="extract-utilities" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.180732 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79f5da6-1ccb-4e02-83ca-772969378d20" containerName="pruner" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180738 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79f5da6-1ccb-4e02-83ca-772969378d20" containerName="pruner" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180787 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180833 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f49b107-bb1a-49c1-ab50-fad36694e25b" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180844 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" containerName="oauth-openshift" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180852 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa58da1b-2409-49a0-a2f3-55e83d584a91" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180863 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1fd2533-7e30-4022-b76b-02a354cdb2d0" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180873 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79f5da6-1ccb-4e02-83ca-772969378d20" containerName="pruner" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.180882 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7970d8-5509-4ec1-a9e0-549db8358313" containerName="registry-server" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.181222 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.183851 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-cr2rf"] Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.192276 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.192316 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc" (OuterVolumeSpecName: "kube-api-access-b4tpc") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "kube-api-access-b4tpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.192558 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.193229 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.194666 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.194925 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.195310 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.205269 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "14cad03b-9a44-4f7f-a3a1-ae8b970fd993" (UID: "14cad03b-9a44-4f7f-a3a1-ae8b970fd993"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.273973 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274033 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274054 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274076 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkscr\" (UniqueName: \"kubernetes.io/projected/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-kube-api-access-bkscr\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274115 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-policies\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274238 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274305 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274338 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274365 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-dir\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274417 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274470 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274492 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274516 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274542 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274620 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274637 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4tpc\" (UniqueName: \"kubernetes.io/projected/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-kube-api-access-b4tpc\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274700 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274717 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274730 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274741 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274754 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274765 4989 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274777 4989 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274790 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274803 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274816 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274844 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.274858 4989 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14cad03b-9a44-4f7f-a3a1-ae8b970fd993-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376340 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376416 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376497 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-dir\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376627 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376731 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376786 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376841 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.376957 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.377022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.377073 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.377139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkscr\" (UniqueName: \"kubernetes.io/projected/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-kube-api-access-bkscr\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.377194 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-policies\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.378213 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-dir\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.378231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.378488 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-audit-policies\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.379886 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-service-ca\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.380181 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.380577 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.380993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.381451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-error\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.381833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-router-certs\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.382333 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.382742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.383001 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-user-template-login\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.383742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-v4-0-config-system-session\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.398924 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkscr\" (UniqueName: \"kubernetes.io/projected/c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab-kube-api-access-bkscr\") pod \"oauth-openshift-5555647bc4-cr2rf\" (UID: \"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab\") " pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.537403 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.763072 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5555647bc4-cr2rf"] Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.910582 4989 generic.go:334] "Generic (PLEG): container finished" podID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" containerID="10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b" exitCode=0 Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.910705 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.910705 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" event={"ID":"14cad03b-9a44-4f7f-a3a1-ae8b970fd993","Type":"ContainerDied","Data":"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b"} Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.910779 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-krvc2" event={"ID":"14cad03b-9a44-4f7f-a3a1-ae8b970fd993","Type":"ContainerDied","Data":"40d16538bdc6e303072d246235af03345f75d4d8ac1cb4f477dab8172b03483a"} Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.910822 4989 scope.go:117] "RemoveContainer" containerID="10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.911873 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" event={"ID":"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab","Type":"ContainerStarted","Data":"6cd1abf5b39365938c3fdac83f2f483c39c3634a7c0a19ab8f8000767652ca91"} Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.957085 4989 scope.go:117] "RemoveContainer" containerID="10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b" Oct 06 08:43:26 crc kubenswrapper[4989]: E1006 08:43:26.957865 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b\": container with ID starting with 10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b not found: ID does not exist" containerID="10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.957915 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b"} err="failed to get container status \"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b\": rpc error: code = NotFound desc = could not find container \"10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b\": container with ID starting with 10c8bedd13cdbc6b2fdd9a546966b5a190f166fd03352fb78b48851a92f2357b not found: ID does not exist" Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.988970 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:43:26 crc kubenswrapper[4989]: I1006 08:43:26.991794 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-krvc2"] Oct 06 08:43:27 crc kubenswrapper[4989]: I1006 08:43:27.920504 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" event={"ID":"c9e5cc77-6bb9-421d-bc76-ec87eb9eb2ab","Type":"ContainerStarted","Data":"04e6fcbdf09de231755b847595f0e66e2152551b5601e75a2b8cc29b0fb3914c"} Oct 06 08:43:27 crc kubenswrapper[4989]: I1006 08:43:27.920901 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:27 crc kubenswrapper[4989]: I1006 08:43:27.926271 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" Oct 06 08:43:27 crc kubenswrapper[4989]: I1006 08:43:27.946320 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14cad03b-9a44-4f7f-a3a1-ae8b970fd993" path="/var/lib/kubelet/pods/14cad03b-9a44-4f7f-a3a1-ae8b970fd993/volumes" Oct 06 08:43:27 crc kubenswrapper[4989]: I1006 08:43:27.986350 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5555647bc4-cr2rf" podStartSLOduration=27.986333111 podStartE2EDuration="27.986333111s" podCreationTimestamp="2025-10-06 08:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:43:27.957565071 +0000 UTC m=+258.747590651" watchObservedRunningTime="2025-10-06 08:43:27.986333111 +0000 UTC m=+258.776358691" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.139397 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc"] Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.140728 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.143927 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.144780 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.152175 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc"] Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.208639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.208779 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7sxx\" (UniqueName: \"kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.208892 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.309560 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.309615 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.309643 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7sxx\" (UniqueName: \"kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.311179 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.322451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.326862 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7sxx\" (UniqueName: \"kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx\") pod \"collect-profiles-29329005-vjcgc\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.466827 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:00 crc kubenswrapper[4989]: I1006 08:45:00.901864 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc"] Oct 06 08:45:01 crc kubenswrapper[4989]: I1006 08:45:01.453966 4989 generic.go:334] "Generic (PLEG): container finished" podID="5dcddcf3-b68d-4327-b441-26cfac414aab" containerID="53ed129a133edb823a9ffcd57f0b3f6e401d72815cdd8ae1f390abea11220dde" exitCode=0 Oct 06 08:45:01 crc kubenswrapper[4989]: I1006 08:45:01.454047 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" event={"ID":"5dcddcf3-b68d-4327-b441-26cfac414aab","Type":"ContainerDied","Data":"53ed129a133edb823a9ffcd57f0b3f6e401d72815cdd8ae1f390abea11220dde"} Oct 06 08:45:01 crc kubenswrapper[4989]: I1006 08:45:01.454092 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" event={"ID":"5dcddcf3-b68d-4327-b441-26cfac414aab","Type":"ContainerStarted","Data":"877f8b0ca9feb9035c27d30a1b6369daba58b999442ea4d7b4f6d76335ca21ff"} Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.694053 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.740498 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume\") pod \"5dcddcf3-b68d-4327-b441-26cfac414aab\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.740574 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7sxx\" (UniqueName: \"kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx\") pod \"5dcddcf3-b68d-4327-b441-26cfac414aab\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.740708 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume\") pod \"5dcddcf3-b68d-4327-b441-26cfac414aab\" (UID: \"5dcddcf3-b68d-4327-b441-26cfac414aab\") " Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.741645 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume" (OuterVolumeSpecName: "config-volume") pod "5dcddcf3-b68d-4327-b441-26cfac414aab" (UID: "5dcddcf3-b68d-4327-b441-26cfac414aab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.746030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx" (OuterVolumeSpecName: "kube-api-access-n7sxx") pod "5dcddcf3-b68d-4327-b441-26cfac414aab" (UID: "5dcddcf3-b68d-4327-b441-26cfac414aab"). InnerVolumeSpecName "kube-api-access-n7sxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.746775 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5dcddcf3-b68d-4327-b441-26cfac414aab" (UID: "5dcddcf3-b68d-4327-b441-26cfac414aab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.841580 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5dcddcf3-b68d-4327-b441-26cfac414aab-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.841631 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5dcddcf3-b68d-4327-b441-26cfac414aab-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:02 crc kubenswrapper[4989]: I1006 08:45:02.841647 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7sxx\" (UniqueName: \"kubernetes.io/projected/5dcddcf3-b68d-4327-b441-26cfac414aab-kube-api-access-n7sxx\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:03 crc kubenswrapper[4989]: I1006 08:45:03.466395 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" event={"ID":"5dcddcf3-b68d-4327-b441-26cfac414aab","Type":"ContainerDied","Data":"877f8b0ca9feb9035c27d30a1b6369daba58b999442ea4d7b4f6d76335ca21ff"} Oct 06 08:45:03 crc kubenswrapper[4989]: I1006 08:45:03.466641 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="877f8b0ca9feb9035c27d30a1b6369daba58b999442ea4d7b4f6d76335ca21ff" Oct 06 08:45:03 crc kubenswrapper[4989]: I1006 08:45:03.466495 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc" Oct 06 08:45:33 crc kubenswrapper[4989]: I1006 08:45:33.935642 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:45:33 crc kubenswrapper[4989]: I1006 08:45:33.936299 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.317439 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.319109 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7xfhm" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="registry-server" containerID="cri-o://c242062bf0407096034796a0010bd22528b184749c4859e0b9792ee6a33f6bf5" gracePeriod=30 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.325399 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.325743 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tnvzq" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="registry-server" containerID="cri-o://e5025a816f27f9acce03351c576f8e7e2baca27fffa094f6d522df42603bb744" gracePeriod=30 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.333287 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.334432 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" containerID="cri-o://6f5821254c80639f2d6c7c794ee6157ce2fabe5e2afb83602f7823008662b330" gracePeriod=30 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.340012 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.350746 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xkxs2"] Oct 06 08:45:42 crc kubenswrapper[4989]: E1006 08:45:42.350998 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcddcf3-b68d-4327-b441-26cfac414aab" containerName="collect-profiles" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.351014 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcddcf3-b68d-4327-b441-26cfac414aab" containerName="collect-profiles" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.351143 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcddcf3-b68d-4327-b441-26cfac414aab" containerName="collect-profiles" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.351633 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.355093 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.355396 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8bfp" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="registry-server" containerID="cri-o://1c24ced7cecfdb737a3624fd3c2d2854604b77c02df3c656884d33efddf2bfb4" gracePeriod=30 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.366696 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xkxs2"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.511122 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdstj\" (UniqueName: \"kubernetes.io/projected/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-kube-api-access-vdstj\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.511225 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.511261 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.612471 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdstj\" (UniqueName: \"kubernetes.io/projected/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-kube-api-access-vdstj\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.612548 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.612574 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.614101 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.618162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.631187 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdstj\" (UniqueName: \"kubernetes.io/projected/fd477c82-fc0b-4d4a-8de4-037f8f30e0e3-kube-api-access-vdstj\") pod \"marketplace-operator-79b997595-xkxs2\" (UID: \"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3\") " pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.671444 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.708199 4989 generic.go:334] "Generic (PLEG): container finished" podID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerID="6f5821254c80639f2d6c7c794ee6157ce2fabe5e2afb83602f7823008662b330" exitCode=0 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.708257 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" event={"ID":"fb18713a-3d90-4486-aa7e-e93ff0718414","Type":"ContainerDied","Data":"6f5821254c80639f2d6c7c794ee6157ce2fabe5e2afb83602f7823008662b330"} Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.709828 4989 generic.go:334] "Generic (PLEG): container finished" podID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerID="1c24ced7cecfdb737a3624fd3c2d2854604b77c02df3c656884d33efddf2bfb4" exitCode=0 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.709864 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerDied","Data":"1c24ced7cecfdb737a3624fd3c2d2854604b77c02df3c656884d33efddf2bfb4"} Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.712320 4989 generic.go:334] "Generic (PLEG): container finished" podID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerID="e5025a816f27f9acce03351c576f8e7e2baca27fffa094f6d522df42603bb744" exitCode=0 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.712379 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerDied","Data":"e5025a816f27f9acce03351c576f8e7e2baca27fffa094f6d522df42603bb744"} Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.715277 4989 generic.go:334] "Generic (PLEG): container finished" podID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerID="c242062bf0407096034796a0010bd22528b184749c4859e0b9792ee6a33f6bf5" exitCode=0 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.715393 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerDied","Data":"c242062bf0407096034796a0010bd22528b184749c4859e0b9792ee6a33f6bf5"} Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.715479 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4q4mm" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="registry-server" containerID="cri-o://107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5" gracePeriod=30 Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.735119 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.799932 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.854346 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.867713 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917389 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content\") pod \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917454 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content\") pod \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917524 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hx4s\" (UniqueName: \"kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s\") pod \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917548 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdwd4\" (UniqueName: \"kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4\") pod \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917567 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities\") pod \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\" (UID: \"092ad7d4-11e3-4399-a9ae-65f548ba3f85\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917647 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities\") pod \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\" (UID: \"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917680 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt5t2\" (UniqueName: \"kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2\") pod \"fb18713a-3d90-4486-aa7e-e93ff0718414\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917696 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl7g2\" (UniqueName: \"kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2\") pod \"e90ef368-0b6b-4c6a-b96f-7530016bc537\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.917721 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca\") pod \"fb18713a-3d90-4486-aa7e-e93ff0718414\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.918291 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "fb18713a-3d90-4486-aa7e-e93ff0718414" (UID: "fb18713a-3d90-4486-aa7e-e93ff0718414"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.924435 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities" (OuterVolumeSpecName: "utilities") pod "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" (UID: "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.926594 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities" (OuterVolumeSpecName: "utilities") pod "092ad7d4-11e3-4399-a9ae-65f548ba3f85" (UID: "092ad7d4-11e3-4399-a9ae-65f548ba3f85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.927910 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2" (OuterVolumeSpecName: "kube-api-access-pt5t2") pod "fb18713a-3d90-4486-aa7e-e93ff0718414" (UID: "fb18713a-3d90-4486-aa7e-e93ff0718414"). InnerVolumeSpecName "kube-api-access-pt5t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.928290 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2" (OuterVolumeSpecName: "kube-api-access-xl7g2") pod "e90ef368-0b6b-4c6a-b96f-7530016bc537" (UID: "e90ef368-0b6b-4c6a-b96f-7530016bc537"). InnerVolumeSpecName "kube-api-access-xl7g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.928369 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4" (OuterVolumeSpecName: "kube-api-access-xdwd4") pod "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" (UID: "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af"). InnerVolumeSpecName "kube-api-access-xdwd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.928996 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s" (OuterVolumeSpecName: "kube-api-access-4hx4s") pod "092ad7d4-11e3-4399-a9ae-65f548ba3f85" (UID: "092ad7d4-11e3-4399-a9ae-65f548ba3f85"). InnerVolumeSpecName "kube-api-access-4hx4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.947864 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xkxs2"] Oct 06 08:45:42 crc kubenswrapper[4989]: I1006 08:45:42.989129 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" (UID: "e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.014235 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "092ad7d4-11e3-4399-a9ae-65f548ba3f85" (UID: "092ad7d4-11e3-4399-a9ae-65f548ba3f85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.018749 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics\") pod \"fb18713a-3d90-4486-aa7e-e93ff0718414\" (UID: \"fb18713a-3d90-4486-aa7e-e93ff0718414\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.018803 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content\") pod \"e90ef368-0b6b-4c6a-b96f-7530016bc537\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.018901 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities\") pod \"e90ef368-0b6b-4c6a-b96f-7530016bc537\" (UID: \"e90ef368-0b6b-4c6a-b96f-7530016bc537\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019191 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hx4s\" (UniqueName: \"kubernetes.io/projected/092ad7d4-11e3-4399-a9ae-65f548ba3f85-kube-api-access-4hx4s\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019222 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdwd4\" (UniqueName: \"kubernetes.io/projected/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-kube-api-access-xdwd4\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019248 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019294 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019430 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt5t2\" (UniqueName: \"kubernetes.io/projected/fb18713a-3d90-4486-aa7e-e93ff0718414-kube-api-access-pt5t2\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019447 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl7g2\" (UniqueName: \"kubernetes.io/projected/e90ef368-0b6b-4c6a-b96f-7530016bc537-kube-api-access-xl7g2\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019459 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019469 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.019499 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/092ad7d4-11e3-4399-a9ae-65f548ba3f85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.020069 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities" (OuterVolumeSpecName: "utilities") pod "e90ef368-0b6b-4c6a-b96f-7530016bc537" (UID: "e90ef368-0b6b-4c6a-b96f-7530016bc537"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.032808 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "fb18713a-3d90-4486-aa7e-e93ff0718414" (UID: "fb18713a-3d90-4486-aa7e-e93ff0718414"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.058301 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.087680 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e90ef368-0b6b-4c6a-b96f-7530016bc537" (UID: "e90ef368-0b6b-4c6a-b96f-7530016bc537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.120410 4989 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/fb18713a-3d90-4486-aa7e-e93ff0718414-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.120449 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.120461 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e90ef368-0b6b-4c6a-b96f-7530016bc537-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.220990 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities\") pod \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.221071 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8xzl\" (UniqueName: \"kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl\") pod \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.221155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content\") pod \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\" (UID: \"92ea6a88-1fac-4940-96e8-ec0a33d9951f\") " Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.221777 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities" (OuterVolumeSpecName: "utilities") pod "92ea6a88-1fac-4940-96e8-ec0a33d9951f" (UID: "92ea6a88-1fac-4940-96e8-ec0a33d9951f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.224078 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl" (OuterVolumeSpecName: "kube-api-access-f8xzl") pod "92ea6a88-1fac-4940-96e8-ec0a33d9951f" (UID: "92ea6a88-1fac-4940-96e8-ec0a33d9951f"). InnerVolumeSpecName "kube-api-access-f8xzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.234060 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92ea6a88-1fac-4940-96e8-ec0a33d9951f" (UID: "92ea6a88-1fac-4940-96e8-ec0a33d9951f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.322220 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.322251 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8xzl\" (UniqueName: \"kubernetes.io/projected/92ea6a88-1fac-4940-96e8-ec0a33d9951f-kube-api-access-f8xzl\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.322261 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92ea6a88-1fac-4940-96e8-ec0a33d9951f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.722516 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8bfp" event={"ID":"092ad7d4-11e3-4399-a9ae-65f548ba3f85","Type":"ContainerDied","Data":"f2a8cde9736d31b401cb081d61f20f3a98d0634b5732622b06c4e457e3cd85fb"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.722573 4989 scope.go:117] "RemoveContainer" containerID="1c24ced7cecfdb737a3624fd3c2d2854604b77c02df3c656884d33efddf2bfb4" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.722710 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8bfp" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.728625 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" event={"ID":"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3","Type":"ContainerStarted","Data":"775074e72e8e1d398dc4de29c51b00b355ccf15befd3e94e315c611297c62c82"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.728749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" event={"ID":"fd477c82-fc0b-4d4a-8de4-037f8f30e0e3","Type":"ContainerStarted","Data":"38a18adc387fdabdca6859c6f0ffcc029f7177f9e77d104770cc43a8dcd3fe03"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.729805 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.733826 4989 generic.go:334] "Generic (PLEG): container finished" podID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerID="107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5" exitCode=0 Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.733884 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q4mm" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.733891 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerDied","Data":"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.733978 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q4mm" event={"ID":"92ea6a88-1fac-4940-96e8-ec0a33d9951f","Type":"ContainerDied","Data":"128200c39e9872dfd2529145a37f39f937c1fe87eff54d43a3af19ce894b01f5"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.735413 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.738504 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvzq" event={"ID":"e90ef368-0b6b-4c6a-b96f-7530016bc537","Type":"ContainerDied","Data":"e946f489117ae219551744bb14b97acd970d8a90911fe2730ad89770b5410be8"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.738543 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvzq" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.740805 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xfhm" event={"ID":"e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af","Type":"ContainerDied","Data":"f98b19b68a80ef374e7dc538da5a485a1c72a8c797f3fba0f83a9df25d2ac34b"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.740840 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xfhm" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.742820 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" event={"ID":"fb18713a-3d90-4486-aa7e-e93ff0718414","Type":"ContainerDied","Data":"2d327f0fd39378fc7981480894414187f52b5a90b1baf0b26016e4eb5d067563"} Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.742865 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-78t58" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.748475 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xkxs2" podStartSLOduration=1.748460339 podStartE2EDuration="1.748460339s" podCreationTimestamp="2025-10-06 08:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:45:43.747107183 +0000 UTC m=+394.537132803" watchObservedRunningTime="2025-10-06 08:45:43.748460339 +0000 UTC m=+394.538485919" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.781066 4989 scope.go:117] "RemoveContainer" containerID="2df10fe20f69a9a91e7d5091a3aa45e18def066b5bb4e7f12f6dd8830dbe1a58" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.800223 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.804727 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8bfp"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.822395 4989 scope.go:117] "RemoveContainer" containerID="c0889d50d9d67470005007e9bb4bdafe51e7ec632b9d9fdcee596fa189478e05" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.823974 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.826340 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q4mm"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.845006 4989 scope.go:117] "RemoveContainer" containerID="107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5" Oct 06 08:45:43 crc kubenswrapper[4989]: E1006 08:45:43.861039 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod092ad7d4_11e3_4399_a9ae_65f548ba3f85.slice/crio-f2a8cde9736d31b401cb081d61f20f3a98d0634b5732622b06c4e457e3cd85fb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92ea6a88_1fac_4940_96e8_ec0a33d9951f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92ea6a88_1fac_4940_96e8_ec0a33d9951f.slice/crio-128200c39e9872dfd2529145a37f39f937c1fe87eff54d43a3af19ce894b01f5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod092ad7d4_11e3_4399_a9ae_65f548ba3f85.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode90ef368_0b6b_4c6a_b96f_7530016bc537.slice\": RecentStats: unable to find data in memory cache]" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.868525 4989 scope.go:117] "RemoveContainer" containerID="38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.871952 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.882078 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7xfhm"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.893453 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.893617 4989 scope.go:117] "RemoveContainer" containerID="a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.896597 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tnvzq"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.899511 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.903243 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-78t58"] Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.907526 4989 scope.go:117] "RemoveContainer" containerID="107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5" Oct 06 08:45:43 crc kubenswrapper[4989]: E1006 08:45:43.908036 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5\": container with ID starting with 107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5 not found: ID does not exist" containerID="107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.908085 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5"} err="failed to get container status \"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5\": rpc error: code = NotFound desc = could not find container \"107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5\": container with ID starting with 107c05e990627bd42749bf8a33758e92e4667751a7e9c58ae687ee773de8d5d5 not found: ID does not exist" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.908119 4989 scope.go:117] "RemoveContainer" containerID="38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633" Oct 06 08:45:43 crc kubenswrapper[4989]: E1006 08:45:43.908559 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633\": container with ID starting with 38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633 not found: ID does not exist" containerID="38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.908694 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633"} err="failed to get container status \"38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633\": rpc error: code = NotFound desc = could not find container \"38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633\": container with ID starting with 38cefb145ab777cb5b76763eaa2a64666c8eca04d12b476fa4c401aa47918633 not found: ID does not exist" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.908795 4989 scope.go:117] "RemoveContainer" containerID="a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee" Oct 06 08:45:43 crc kubenswrapper[4989]: E1006 08:45:43.909295 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee\": container with ID starting with a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee not found: ID does not exist" containerID="a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.909332 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee"} err="failed to get container status \"a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee\": rpc error: code = NotFound desc = could not find container \"a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee\": container with ID starting with a926aa1c772f17c4e96d473967db1a6ef68db78ee7b743f4c388ce4396d20fee not found: ID does not exist" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.909358 4989 scope.go:117] "RemoveContainer" containerID="e5025a816f27f9acce03351c576f8e7e2baca27fffa094f6d522df42603bb744" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.922956 4989 scope.go:117] "RemoveContainer" containerID="d703d13a9adb8cc784c416b62c5901c76d0c805e2fa0a339689d98d8c40d5f8b" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.939898 4989 scope.go:117] "RemoveContainer" containerID="af8fbf4c6bd23d32f4999b8714b7549353870f83d7579e049bbf53f921a862bb" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.942978 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" path="/var/lib/kubelet/pods/092ad7d4-11e3-4399-a9ae-65f548ba3f85/volumes" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.944094 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" path="/var/lib/kubelet/pods/92ea6a88-1fac-4940-96e8-ec0a33d9951f/volumes" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.944982 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" path="/var/lib/kubelet/pods/e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af/volumes" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.945922 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" path="/var/lib/kubelet/pods/e90ef368-0b6b-4c6a-b96f-7530016bc537/volumes" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.947486 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" path="/var/lib/kubelet/pods/fb18713a-3d90-4486-aa7e-e93ff0718414/volumes" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.952872 4989 scope.go:117] "RemoveContainer" containerID="c242062bf0407096034796a0010bd22528b184749c4859e0b9792ee6a33f6bf5" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.966260 4989 scope.go:117] "RemoveContainer" containerID="91ff25207f07f96ec836fae08a23e60f7c522bb36f807193a86b1ac58b3d8682" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.983020 4989 scope.go:117] "RemoveContainer" containerID="7a911560c3e865cba1f7875fbb1fc46fe7f621d571b166e361517971d51a95be" Oct 06 08:45:43 crc kubenswrapper[4989]: I1006 08:45:43.998979 4989 scope.go:117] "RemoveContainer" containerID="6f5821254c80639f2d6c7c794ee6157ce2fabe5e2afb83602f7823008662b330" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528017 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vxfgz"] Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528246 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528261 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528273 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528281 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528292 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528300 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528309 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528317 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528333 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528340 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="extract-utilities" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528349 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528356 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528365 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528373 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528384 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528391 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528403 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528410 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528419 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528428 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528438 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528447 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528461 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528469 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="extract-content" Oct 06 08:45:44 crc kubenswrapper[4989]: E1006 08:45:44.528478 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528484 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528582 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="092ad7d4-11e3-4399-a9ae-65f548ba3f85" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528598 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e6cc6c-2cc4-4888-af28-c2f7fd02f3af" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528608 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ea6a88-1fac-4940-96e8-ec0a33d9951f" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528618 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e90ef368-0b6b-4c6a-b96f-7530016bc537" containerName="registry-server" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.528626 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb18713a-3d90-4486-aa7e-e93ff0718414" containerName="marketplace-operator" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.529644 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.531782 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.536913 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-catalog-content\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.537135 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-utilities\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.537338 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc6c4\" (UniqueName: \"kubernetes.io/projected/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-kube-api-access-kc6c4\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.539097 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxfgz"] Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.638327 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc6c4\" (UniqueName: \"kubernetes.io/projected/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-kube-api-access-kc6c4\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.638389 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-catalog-content\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.638423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-utilities\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.639163 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-utilities\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.639260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-catalog-content\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.657255 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc6c4\" (UniqueName: \"kubernetes.io/projected/50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf-kube-api-access-kc6c4\") pod \"redhat-marketplace-vxfgz\" (UID: \"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf\") " pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.731944 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cjkm6"] Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.733327 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.736044 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.738937 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-utilities\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.738976 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-catalog-content\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.739029 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpvs6\" (UniqueName: \"kubernetes.io/projected/aaae90ae-1c87-4233-85aa-98d3dbbc7790-kube-api-access-dpvs6\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.741002 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjkm6"] Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.839910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpvs6\" (UniqueName: \"kubernetes.io/projected/aaae90ae-1c87-4233-85aa-98d3dbbc7790-kube-api-access-dpvs6\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.840300 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-utilities\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.840331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-catalog-content\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.841076 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-utilities\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.841115 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaae90ae-1c87-4233-85aa-98d3dbbc7790-catalog-content\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.854681 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:44 crc kubenswrapper[4989]: I1006 08:45:44.857297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpvs6\" (UniqueName: \"kubernetes.io/projected/aaae90ae-1c87-4233-85aa-98d3dbbc7790-kube-api-access-dpvs6\") pod \"community-operators-cjkm6\" (UID: \"aaae90ae-1c87-4233-85aa-98d3dbbc7790\") " pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.055195 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.230103 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjkm6"] Oct 06 08:45:45 crc kubenswrapper[4989]: W1006 08:45:45.231346 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50b0a5d8_8aa4_45e4_bf00_c6e717c3c1bf.slice/crio-3d8f9de6cf82907410888458971992d99879c2aec0d3c835e3eb28f99eb555f9 WatchSource:0}: Error finding container 3d8f9de6cf82907410888458971992d99879c2aec0d3c835e3eb28f99eb555f9: Status 404 returned error can't find the container with id 3d8f9de6cf82907410888458971992d99879c2aec0d3c835e3eb28f99eb555f9 Oct 06 08:45:45 crc kubenswrapper[4989]: W1006 08:45:45.232046 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaae90ae_1c87_4233_85aa_98d3dbbc7790.slice/crio-ffa2a41b9b023064a83d4da248e4ed1646e660dfcd7cb83fe7daf517791c5ee9 WatchSource:0}: Error finding container ffa2a41b9b023064a83d4da248e4ed1646e660dfcd7cb83fe7daf517791c5ee9: Status 404 returned error can't find the container with id ffa2a41b9b023064a83d4da248e4ed1646e660dfcd7cb83fe7daf517791c5ee9 Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.233399 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxfgz"] Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.762577 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaae90ae-1c87-4233-85aa-98d3dbbc7790" containerID="3e926de7ecc2ab017d919a6661fcf82284cc5ef8c936f6b3ff95bdb96941c1bd" exitCode=0 Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.762623 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjkm6" event={"ID":"aaae90ae-1c87-4233-85aa-98d3dbbc7790","Type":"ContainerDied","Data":"3e926de7ecc2ab017d919a6661fcf82284cc5ef8c936f6b3ff95bdb96941c1bd"} Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.762964 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjkm6" event={"ID":"aaae90ae-1c87-4233-85aa-98d3dbbc7790","Type":"ContainerStarted","Data":"ffa2a41b9b023064a83d4da248e4ed1646e660dfcd7cb83fe7daf517791c5ee9"} Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.766165 4989 generic.go:334] "Generic (PLEG): container finished" podID="50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf" containerID="72687e9c7a7bb6cea73516e2eba70a672aad52b5db7113949c368ef0b8e47297" exitCode=0 Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.766937 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxfgz" event={"ID":"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf","Type":"ContainerDied","Data":"72687e9c7a7bb6cea73516e2eba70a672aad52b5db7113949c368ef0b8e47297"} Oct 06 08:45:45 crc kubenswrapper[4989]: I1006 08:45:45.766968 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxfgz" event={"ID":"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf","Type":"ContainerStarted","Data":"3d8f9de6cf82907410888458971992d99879c2aec0d3c835e3eb28f99eb555f9"} Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.929466 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pwzh2"] Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.933586 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.935981 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.943017 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pwzh2"] Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.965157 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-utilities\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.965336 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72szg\" (UniqueName: \"kubernetes.io/projected/0b11d153-4f54-4a5e-9a21-69527194a1e4-kube-api-access-72szg\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:46 crc kubenswrapper[4989]: I1006 08:45:46.965491 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-catalog-content\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.066133 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-catalog-content\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.066242 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-utilities\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.066272 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72szg\" (UniqueName: \"kubernetes.io/projected/0b11d153-4f54-4a5e-9a21-69527194a1e4-kube-api-access-72szg\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.068437 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-utilities\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.078320 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b11d153-4f54-4a5e-9a21-69527194a1e4-catalog-content\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.084575 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72szg\" (UniqueName: \"kubernetes.io/projected/0b11d153-4f54-4a5e-9a21-69527194a1e4-kube-api-access-72szg\") pod \"certified-operators-pwzh2\" (UID: \"0b11d153-4f54-4a5e-9a21-69527194a1e4\") " pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.130340 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.131636 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.133666 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.140663 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.167132 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nmpg\" (UniqueName: \"kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.167194 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.167216 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.258917 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.269711 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.270000 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.270117 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nmpg\" (UniqueName: \"kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.270238 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.270487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.288559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nmpg\" (UniqueName: \"kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg\") pod \"redhat-operators-85sl7\" (UID: \"796add64-e739-490b-862c-274297646887\") " pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.436159 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pwzh2"] Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.460391 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.643513 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 08:45:47 crc kubenswrapper[4989]: W1006 08:45:47.652506 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod796add64_e739_490b_862c_274297646887.slice/crio-98dc2942d603462b5f61bf909447d59c7f2204a6c9c305ee15db4b473285ff17 WatchSource:0}: Error finding container 98dc2942d603462b5f61bf909447d59c7f2204a6c9c305ee15db4b473285ff17: Status 404 returned error can't find the container with id 98dc2942d603462b5f61bf909447d59c7f2204a6c9c305ee15db4b473285ff17 Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.781459 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerStarted","Data":"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495"} Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.781501 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerStarted","Data":"98dc2942d603462b5f61bf909447d59c7f2204a6c9c305ee15db4b473285ff17"} Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.782504 4989 generic.go:334] "Generic (PLEG): container finished" podID="0b11d153-4f54-4a5e-9a21-69527194a1e4" containerID="b56fda4f3936b3a83624e6e5d54480f284f7d4921a8caa709e31a13cd2bafaf9" exitCode=0 Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.782544 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwzh2" event={"ID":"0b11d153-4f54-4a5e-9a21-69527194a1e4","Type":"ContainerDied","Data":"b56fda4f3936b3a83624e6e5d54480f284f7d4921a8caa709e31a13cd2bafaf9"} Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.782589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwzh2" event={"ID":"0b11d153-4f54-4a5e-9a21-69527194a1e4","Type":"ContainerStarted","Data":"c90a5834beba18b40ef0fc6432dee61a37d517ce82ef8fda7a5981133faced5b"} Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.785805 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaae90ae-1c87-4233-85aa-98d3dbbc7790" containerID="ecdcff132d14edf0fa8e4c4a206d7fe39a4c374955fdeb48b2005c703d7b04e9" exitCode=0 Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.785855 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjkm6" event={"ID":"aaae90ae-1c87-4233-85aa-98d3dbbc7790","Type":"ContainerDied","Data":"ecdcff132d14edf0fa8e4c4a206d7fe39a4c374955fdeb48b2005c703d7b04e9"} Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.788567 4989 generic.go:334] "Generic (PLEG): container finished" podID="50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf" containerID="e04477f292585fbe58c8d2262eb5280fc8df253cdbd982ad75bacfc281bf880e" exitCode=0 Oct 06 08:45:47 crc kubenswrapper[4989]: I1006 08:45:47.788589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxfgz" event={"ID":"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf","Type":"ContainerDied","Data":"e04477f292585fbe58c8d2262eb5280fc8df253cdbd982ad75bacfc281bf880e"} Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.796364 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjkm6" event={"ID":"aaae90ae-1c87-4233-85aa-98d3dbbc7790","Type":"ContainerStarted","Data":"f6969d466858d7d60342916d3d929b7f512779f0acbad13a141f26c5b784a94b"} Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.798933 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxfgz" event={"ID":"50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf","Type":"ContainerStarted","Data":"760daf6b48c121834f1326fbcdf1d4a653980d2c16070098eee7294dfd9ecd12"} Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.802322 4989 generic.go:334] "Generic (PLEG): container finished" podID="796add64-e739-490b-862c-274297646887" containerID="41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495" exitCode=0 Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.802372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerDied","Data":"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495"} Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.820204 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cjkm6" podStartSLOduration=2.113211282 podStartE2EDuration="4.82018592s" podCreationTimestamp="2025-10-06 08:45:44 +0000 UTC" firstStartedPulling="2025-10-06 08:45:45.764190431 +0000 UTC m=+396.554216021" lastFinishedPulling="2025-10-06 08:45:48.471165079 +0000 UTC m=+399.261190659" observedRunningTime="2025-10-06 08:45:48.817534308 +0000 UTC m=+399.607559888" watchObservedRunningTime="2025-10-06 08:45:48.82018592 +0000 UTC m=+399.610211500" Oct 06 08:45:48 crc kubenswrapper[4989]: I1006 08:45:48.849663 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vxfgz" podStartSLOduration=2.363934184 podStartE2EDuration="4.849632123s" podCreationTimestamp="2025-10-06 08:45:44 +0000 UTC" firstStartedPulling="2025-10-06 08:45:45.767997094 +0000 UTC m=+396.558022674" lastFinishedPulling="2025-10-06 08:45:48.253695033 +0000 UTC m=+399.043720613" observedRunningTime="2025-10-06 08:45:48.848154782 +0000 UTC m=+399.638180372" watchObservedRunningTime="2025-10-06 08:45:48.849632123 +0000 UTC m=+399.639657703" Oct 06 08:45:49 crc kubenswrapper[4989]: I1006 08:45:49.808558 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerStarted","Data":"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae"} Oct 06 08:45:49 crc kubenswrapper[4989]: I1006 08:45:49.810253 4989 generic.go:334] "Generic (PLEG): container finished" podID="0b11d153-4f54-4a5e-9a21-69527194a1e4" containerID="7edf4cc5329de013b358042aaab7d545da6d6754f5e42e91e4ce29d9c63bb147" exitCode=0 Oct 06 08:45:49 crc kubenswrapper[4989]: I1006 08:45:49.810292 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwzh2" event={"ID":"0b11d153-4f54-4a5e-9a21-69527194a1e4","Type":"ContainerDied","Data":"7edf4cc5329de013b358042aaab7d545da6d6754f5e42e91e4ce29d9c63bb147"} Oct 06 08:45:50 crc kubenswrapper[4989]: I1006 08:45:50.817542 4989 generic.go:334] "Generic (PLEG): container finished" podID="796add64-e739-490b-862c-274297646887" containerID="dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae" exitCode=0 Oct 06 08:45:50 crc kubenswrapper[4989]: I1006 08:45:50.817623 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerDied","Data":"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae"} Oct 06 08:45:51 crc kubenswrapper[4989]: I1006 08:45:51.825857 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pwzh2" event={"ID":"0b11d153-4f54-4a5e-9a21-69527194a1e4","Type":"ContainerStarted","Data":"53e93549446c29a0dbaa0ada2720c1ad81624cfc3c81c2a1f7fbcc3f30829444"} Oct 06 08:45:51 crc kubenswrapper[4989]: I1006 08:45:51.845754 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pwzh2" podStartSLOduration=3.296553133 podStartE2EDuration="5.845734941s" podCreationTimestamp="2025-10-06 08:45:46 +0000 UTC" firstStartedPulling="2025-10-06 08:45:47.784175168 +0000 UTC m=+398.574200738" lastFinishedPulling="2025-10-06 08:45:50.333356956 +0000 UTC m=+401.123382546" observedRunningTime="2025-10-06 08:45:51.842793721 +0000 UTC m=+402.632819311" watchObservedRunningTime="2025-10-06 08:45:51.845734941 +0000 UTC m=+402.635760531" Oct 06 08:45:53 crc kubenswrapper[4989]: I1006 08:45:53.835940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerStarted","Data":"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd"} Oct 06 08:45:53 crc kubenswrapper[4989]: I1006 08:45:53.857715 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-85sl7" podStartSLOduration=2.740744796 podStartE2EDuration="6.857700319s" podCreationTimestamp="2025-10-06 08:45:47 +0000 UTC" firstStartedPulling="2025-10-06 08:45:48.825223417 +0000 UTC m=+399.615248997" lastFinishedPulling="2025-10-06 08:45:52.94217892 +0000 UTC m=+403.732204520" observedRunningTime="2025-10-06 08:45:53.856094105 +0000 UTC m=+404.646119705" watchObservedRunningTime="2025-10-06 08:45:53.857700319 +0000 UTC m=+404.647725899" Oct 06 08:45:54 crc kubenswrapper[4989]: I1006 08:45:54.855335 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:54 crc kubenswrapper[4989]: I1006 08:45:54.855689 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:54 crc kubenswrapper[4989]: I1006 08:45:54.912802 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:55 crc kubenswrapper[4989]: I1006 08:45:55.056023 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:55 crc kubenswrapper[4989]: I1006 08:45:55.056072 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:55 crc kubenswrapper[4989]: I1006 08:45:55.097316 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:55 crc kubenswrapper[4989]: I1006 08:45:55.892368 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cjkm6" Oct 06 08:45:55 crc kubenswrapper[4989]: I1006 08:45:55.897354 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vxfgz" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.260045 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.260303 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.310442 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.461117 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.461163 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.507503 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:45:57 crc kubenswrapper[4989]: I1006 08:45:57.902100 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pwzh2" Oct 06 08:46:03 crc kubenswrapper[4989]: I1006 08:46:03.935521 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:46:03 crc kubenswrapper[4989]: I1006 08:46:03.935847 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:46:07 crc kubenswrapper[4989]: I1006 08:46:07.506254 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 08:46:33 crc kubenswrapper[4989]: I1006 08:46:33.935996 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:46:33 crc kubenswrapper[4989]: I1006 08:46:33.936760 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:46:33 crc kubenswrapper[4989]: I1006 08:46:33.942012 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:46:33 crc kubenswrapper[4989]: I1006 08:46:33.942606 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 08:46:33 crc kubenswrapper[4989]: I1006 08:46:33.942692 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97" gracePeriod=600 Oct 06 08:46:35 crc kubenswrapper[4989]: I1006 08:46:35.054430 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97" exitCode=0 Oct 06 08:46:35 crc kubenswrapper[4989]: I1006 08:46:35.054508 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97"} Oct 06 08:46:35 crc kubenswrapper[4989]: I1006 08:46:35.055014 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2"} Oct 06 08:46:35 crc kubenswrapper[4989]: I1006 08:46:35.055038 4989 scope.go:117] "RemoveContainer" containerID="e7ce5ad203775a467be1879c1c0af19cfe5cb3a786dc52531c2e285cb1eccd68" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.080210 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqnt5"] Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.081219 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.100987 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqnt5"] Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252686 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252728 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd53f875-6940-440b-a1ce-5dd02af1e77e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252752 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8dw6\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-kube-api-access-s8dw6\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252781 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-trusted-ca\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd53f875-6940-440b-a1ce-5dd02af1e77e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252833 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-certificates\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252863 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-tls\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.252883 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-bound-sa-token\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.272854 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd53f875-6940-440b-a1ce-5dd02af1e77e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354516 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8dw6\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-kube-api-access-s8dw6\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354568 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-trusted-ca\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd53f875-6940-440b-a1ce-5dd02af1e77e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-certificates\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354724 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-tls\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.354745 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-bound-sa-token\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.355635 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cd53f875-6940-440b-a1ce-5dd02af1e77e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.355984 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-certificates\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.356507 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cd53f875-6940-440b-a1ce-5dd02af1e77e-trusted-ca\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.368719 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-registry-tls\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.368912 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cd53f875-6940-440b-a1ce-5dd02af1e77e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.374015 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8dw6\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-kube-api-access-s8dw6\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.388528 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cd53f875-6940-440b-a1ce-5dd02af1e77e-bound-sa-token\") pod \"image-registry-66df7c8f76-wqnt5\" (UID: \"cd53f875-6940-440b-a1ce-5dd02af1e77e\") " pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.400559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:45 crc kubenswrapper[4989]: I1006 08:47:45.595606 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wqnt5"] Oct 06 08:47:46 crc kubenswrapper[4989]: I1006 08:47:46.449561 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" event={"ID":"cd53f875-6940-440b-a1ce-5dd02af1e77e","Type":"ContainerStarted","Data":"843fefaf23c3c6cbbd5a217c4ce66977018554b48611c8bd62e968d93f392432"} Oct 06 08:47:46 crc kubenswrapper[4989]: I1006 08:47:46.449884 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:47:46 crc kubenswrapper[4989]: I1006 08:47:46.449895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" event={"ID":"cd53f875-6940-440b-a1ce-5dd02af1e77e","Type":"ContainerStarted","Data":"1b6a44fffc87cd551b3b515c70e124d1e0f64d9ce0474de92e4441d211e83097"} Oct 06 08:47:46 crc kubenswrapper[4989]: I1006 08:47:46.468243 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" podStartSLOduration=1.468225717 podStartE2EDuration="1.468225717s" podCreationTimestamp="2025-10-06 08:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:47:46.465440363 +0000 UTC m=+517.255465953" watchObservedRunningTime="2025-10-06 08:47:46.468225717 +0000 UTC m=+517.258251297" Oct 06 08:48:05 crc kubenswrapper[4989]: I1006 08:48:05.405925 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wqnt5" Oct 06 08:48:05 crc kubenswrapper[4989]: I1006 08:48:05.467793 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.510002 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" podUID="88a26819-2d38-493d-8792-747304805368" containerName="registry" containerID="cri-o://f90da1cbebdf01715c0b456773a23ccf4928182ddde59d2805d8dcb3a0c24fb3" gracePeriod=30 Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.713431 4989 generic.go:334] "Generic (PLEG): container finished" podID="88a26819-2d38-493d-8792-747304805368" containerID="f90da1cbebdf01715c0b456773a23ccf4928182ddde59d2805d8dcb3a0c24fb3" exitCode=0 Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.713476 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" event={"ID":"88a26819-2d38-493d-8792-747304805368","Type":"ContainerDied","Data":"f90da1cbebdf01715c0b456773a23ccf4928182ddde59d2805d8dcb3a0c24fb3"} Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.842721 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970476 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970526 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970568 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970606 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970755 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970776 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbz42\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.970795 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls\") pod \"88a26819-2d38-493d-8792-747304805368\" (UID: \"88a26819-2d38-493d-8792-747304805368\") " Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.971749 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.971920 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.978472 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.979371 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.979461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42" (OuterVolumeSpecName: "kube-api-access-dbz42") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "kube-api-access-dbz42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.983018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.985219 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 08:48:30 crc kubenswrapper[4989]: I1006 08:48:30.986076 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "88a26819-2d38-493d-8792-747304805368" (UID: "88a26819-2d38-493d-8792-747304805368"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072301 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072337 4989 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/88a26819-2d38-493d-8792-747304805368-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072350 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbz42\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-kube-api-access-dbz42\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072361 4989 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072372 4989 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/88a26819-2d38-493d-8792-747304805368-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072383 4989 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/88a26819-2d38-493d-8792-747304805368-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.072394 4989 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/88a26819-2d38-493d-8792-747304805368-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.720804 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" event={"ID":"88a26819-2d38-493d-8792-747304805368","Type":"ContainerDied","Data":"4b88b6d4dbde3aa37c2c257e8fe5214f7f2a57cd0b74f9eb48f276f63859cc9f"} Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.720878 4989 scope.go:117] "RemoveContainer" containerID="f90da1cbebdf01715c0b456773a23ccf4928182ddde59d2805d8dcb3a0c24fb3" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.720904 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p8vh6" Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.756162 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.759862 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p8vh6"] Oct 06 08:48:31 crc kubenswrapper[4989]: I1006 08:48:31.942008 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a26819-2d38-493d-8792-747304805368" path="/var/lib/kubelet/pods/88a26819-2d38-493d-8792-747304805368/volumes" Oct 06 08:49:03 crc kubenswrapper[4989]: I1006 08:49:03.936007 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:49:03 crc kubenswrapper[4989]: I1006 08:49:03.936686 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:49:33 crc kubenswrapper[4989]: I1006 08:49:33.935484 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:49:33 crc kubenswrapper[4989]: I1006 08:49:33.936107 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:50:03 crc kubenswrapper[4989]: I1006 08:50:03.935127 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:50:03 crc kubenswrapper[4989]: I1006 08:50:03.937658 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:50:03 crc kubenswrapper[4989]: I1006 08:50:03.951446 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:50:03 crc kubenswrapper[4989]: I1006 08:50:03.952077 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 08:50:03 crc kubenswrapper[4989]: I1006 08:50:03.952155 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2" gracePeriod=600 Oct 06 08:50:04 crc kubenswrapper[4989]: I1006 08:50:04.234173 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2" exitCode=0 Oct 06 08:50:04 crc kubenswrapper[4989]: I1006 08:50:04.234273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2"} Oct 06 08:50:04 crc kubenswrapper[4989]: I1006 08:50:04.234516 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a"} Oct 06 08:50:04 crc kubenswrapper[4989]: I1006 08:50:04.234546 4989 scope.go:117] "RemoveContainer" containerID="5be631c11e927ecaea1f586c3eedc85bb61654333d27434b775c15e870c25c97" Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.673840 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.674685 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" containerName="controller-manager" containerID="cri-o://f901633d593147837b285f836787e385dc23758abb8512b4b8b5d21caea4fbdb" gracePeriod=30 Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.764361 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.764585 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerName="route-controller-manager" containerID="cri-o://c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458" gracePeriod=30 Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.834008 4989 generic.go:334] "Generic (PLEG): container finished" podID="2b4d3115-9287-41a4-9435-1d34590a5178" containerID="f901633d593147837b285f836787e385dc23758abb8512b4b8b5d21caea4fbdb" exitCode=0 Oct 06 08:51:47 crc kubenswrapper[4989]: I1006 08:51:47.834047 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" event={"ID":"2b4d3115-9287-41a4-9435-1d34590a5178","Type":"ContainerDied","Data":"f901633d593147837b285f836787e385dc23758abb8512b4b8b5d21caea4fbdb"} Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.033397 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.049875 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2b4d3115-9287-41a4-9435-1d34590a5178" (UID: "2b4d3115-9287-41a4-9435-1d34590a5178"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.048128 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles\") pod \"2b4d3115-9287-41a4-9435-1d34590a5178\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.053151 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca\") pod \"2b4d3115-9287-41a4-9435-1d34590a5178\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.053184 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcshv\" (UniqueName: \"kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv\") pod \"2b4d3115-9287-41a4-9435-1d34590a5178\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.053214 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert\") pod \"2b4d3115-9287-41a4-9435-1d34590a5178\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.053292 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config\") pod \"2b4d3115-9287-41a4-9435-1d34590a5178\" (UID: \"2b4d3115-9287-41a4-9435-1d34590a5178\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.053567 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.054038 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config" (OuterVolumeSpecName: "config") pod "2b4d3115-9287-41a4-9435-1d34590a5178" (UID: "2b4d3115-9287-41a4-9435-1d34590a5178"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.054457 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca" (OuterVolumeSpecName: "client-ca") pod "2b4d3115-9287-41a4-9435-1d34590a5178" (UID: "2b4d3115-9287-41a4-9435-1d34590a5178"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.073383 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv" (OuterVolumeSpecName: "kube-api-access-gcshv") pod "2b4d3115-9287-41a4-9435-1d34590a5178" (UID: "2b4d3115-9287-41a4-9435-1d34590a5178"). InnerVolumeSpecName "kube-api-access-gcshv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.074698 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2b4d3115-9287-41a4-9435-1d34590a5178" (UID: "2b4d3115-9287-41a4-9435-1d34590a5178"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.103204 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154512 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca\") pod \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config\") pod \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154733 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert\") pod \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154787 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ns7v\" (UniqueName: \"kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v\") pod \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\" (UID: \"c8c4448a-914d-4e2c-9ad7-71a2d0444669\") " Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154975 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154990 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b4d3115-9287-41a4-9435-1d34590a5178-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.154999 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcshv\" (UniqueName: \"kubernetes.io/projected/2b4d3115-9287-41a4-9435-1d34590a5178-kube-api-access-gcshv\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.155008 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b4d3115-9287-41a4-9435-1d34590a5178-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.155366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config" (OuterVolumeSpecName: "config") pod "c8c4448a-914d-4e2c-9ad7-71a2d0444669" (UID: "c8c4448a-914d-4e2c-9ad7-71a2d0444669"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.155362 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca" (OuterVolumeSpecName: "client-ca") pod "c8c4448a-914d-4e2c-9ad7-71a2d0444669" (UID: "c8c4448a-914d-4e2c-9ad7-71a2d0444669"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.159136 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c8c4448a-914d-4e2c-9ad7-71a2d0444669" (UID: "c8c4448a-914d-4e2c-9ad7-71a2d0444669"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.162068 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v" (OuterVolumeSpecName: "kube-api-access-2ns7v") pod "c8c4448a-914d-4e2c-9ad7-71a2d0444669" (UID: "c8c4448a-914d-4e2c-9ad7-71a2d0444669"). InnerVolumeSpecName "kube-api-access-2ns7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.256018 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ns7v\" (UniqueName: \"kubernetes.io/projected/c8c4448a-914d-4e2c-9ad7-71a2d0444669-kube-api-access-2ns7v\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.256066 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.256079 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8c4448a-914d-4e2c-9ad7-71a2d0444669-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.256091 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8c4448a-914d-4e2c-9ad7-71a2d0444669-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.839785 4989 generic.go:334] "Generic (PLEG): container finished" podID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerID="c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458" exitCode=0 Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.839832 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.839826 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" event={"ID":"c8c4448a-914d-4e2c-9ad7-71a2d0444669","Type":"ContainerDied","Data":"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458"} Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.841065 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn" event={"ID":"c8c4448a-914d-4e2c-9ad7-71a2d0444669","Type":"ContainerDied","Data":"6b896d6433e032207102d589e22e13a354109859756280afe16aa745e631a9b7"} Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.841102 4989 scope.go:117] "RemoveContainer" containerID="c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.842952 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" event={"ID":"2b4d3115-9287-41a4-9435-1d34590a5178","Type":"ContainerDied","Data":"08c80de1a17b90f7d531aba4cfccfa6e6ee64b8d171eb3ed69b8d159bb333775"} Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.843064 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vhcsz" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.869766 4989 scope.go:117] "RemoveContainer" containerID="c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458" Oct 06 08:51:48 crc kubenswrapper[4989]: E1006 08:51:48.870276 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458\": container with ID starting with c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458 not found: ID does not exist" containerID="c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.870324 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458"} err="failed to get container status \"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458\": rpc error: code = NotFound desc = could not find container \"c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458\": container with ID starting with c4368b1a85f9fde41d1fd10943f542aebf9fab5a0925c19980e32f47ec84a458 not found: ID does not exist" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.870353 4989 scope.go:117] "RemoveContainer" containerID="f901633d593147837b285f836787e385dc23758abb8512b4b8b5d21caea4fbdb" Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.891435 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.898542 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ww9xn"] Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.904531 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:51:48 crc kubenswrapper[4989]: I1006 08:51:48.910304 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vhcsz"] Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.012831 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf"] Oct 06 08:51:49 crc kubenswrapper[4989]: E1006 08:51:49.013145 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerName="route-controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013168 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerName="route-controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: E1006 08:51:49.013178 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a26819-2d38-493d-8792-747304805368" containerName="registry" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013186 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a26819-2d38-493d-8792-747304805368" containerName="registry" Oct 06 08:51:49 crc kubenswrapper[4989]: E1006 08:51:49.013201 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" containerName="controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013209 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" containerName="controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013359 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" containerName="controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013376 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" containerName="route-controller-manager" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013387 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a26819-2d38-493d-8792-747304805368" containerName="registry" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.013846 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.015990 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cf875bb9f-wsn4m"] Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.016332 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.016535 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.016782 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.016952 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.017059 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.017251 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.017340 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.019916 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.020525 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.020537 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.021246 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.021404 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.021428 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.029108 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.039417 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cf875bb9f-wsn4m"] Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.042005 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf"] Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071500 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071568 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071589 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071627 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071683 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071711 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071737 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071839 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5j4x\" (UniqueName: \"kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.071946 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkf8k\" (UniqueName: \"kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.077641 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cf875bb9f-wsn4m"] Oct 06 08:51:49 crc kubenswrapper[4989]: E1006 08:51:49.078035 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-bkf8k proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" podUID="fcfaeace-bfb8-457f-8817-c4404d51e37d" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.098305 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf"] Oct 06 08:51:49 crc kubenswrapper[4989]: E1006 08:51:49.098715 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-r5j4x serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" podUID="61b0f49c-ce6e-4dd2-be73-75451006164f" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173606 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173630 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173674 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5j4x\" (UniqueName: \"kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173711 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkf8k\" (UniqueName: \"kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173757 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173796 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173820 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.173851 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.174782 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.175033 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.175146 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.175598 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.175648 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.179079 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.180826 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.209626 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkf8k\" (UniqueName: \"kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k\") pod \"controller-manager-cf875bb9f-wsn4m\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.210613 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5j4x\" (UniqueName: \"kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x\") pod \"route-controller-manager-677bf5f74d-2bvcf\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.849201 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.849225 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.857203 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.862801 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882007 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert\") pod \"61b0f49c-ce6e-4dd2-be73-75451006164f\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882141 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config\") pod \"fcfaeace-bfb8-457f-8817-c4404d51e37d\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882198 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles\") pod \"fcfaeace-bfb8-457f-8817-c4404d51e37d\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca\") pod \"fcfaeace-bfb8-457f-8817-c4404d51e37d\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882324 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert\") pod \"fcfaeace-bfb8-457f-8817-c4404d51e37d\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882356 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca\") pod \"61b0f49c-ce6e-4dd2-be73-75451006164f\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882388 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkf8k\" (UniqueName: \"kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k\") pod \"fcfaeace-bfb8-457f-8817-c4404d51e37d\" (UID: \"fcfaeace-bfb8-457f-8817-c4404d51e37d\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882438 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config\") pod \"61b0f49c-ce6e-4dd2-be73-75451006164f\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882495 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5j4x\" (UniqueName: \"kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x\") pod \"61b0f49c-ce6e-4dd2-be73-75451006164f\" (UID: \"61b0f49c-ce6e-4dd2-be73-75451006164f\") " Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882760 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config" (OuterVolumeSpecName: "config") pod "fcfaeace-bfb8-457f-8817-c4404d51e37d" (UID: "fcfaeace-bfb8-457f-8817-c4404d51e37d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882713 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "fcfaeace-bfb8-457f-8817-c4404d51e37d" (UID: "fcfaeace-bfb8-457f-8817-c4404d51e37d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882950 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca" (OuterVolumeSpecName: "client-ca") pod "fcfaeace-bfb8-457f-8817-c4404d51e37d" (UID: "fcfaeace-bfb8-457f-8817-c4404d51e37d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.882957 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca" (OuterVolumeSpecName: "client-ca") pod "61b0f49c-ce6e-4dd2-be73-75451006164f" (UID: "61b0f49c-ce6e-4dd2-be73-75451006164f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.883520 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config" (OuterVolumeSpecName: "config") pod "61b0f49c-ce6e-4dd2-be73-75451006164f" (UID: "61b0f49c-ce6e-4dd2-be73-75451006164f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.884325 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.884366 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.884384 4989 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.884404 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fcfaeace-bfb8-457f-8817-c4404d51e37d-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.884420 4989 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61b0f49c-ce6e-4dd2-be73-75451006164f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.886682 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x" (OuterVolumeSpecName: "kube-api-access-r5j4x") pod "61b0f49c-ce6e-4dd2-be73-75451006164f" (UID: "61b0f49c-ce6e-4dd2-be73-75451006164f"). InnerVolumeSpecName "kube-api-access-r5j4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.887049 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "61b0f49c-ce6e-4dd2-be73-75451006164f" (UID: "61b0f49c-ce6e-4dd2-be73-75451006164f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.887073 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k" (OuterVolumeSpecName: "kube-api-access-bkf8k") pod "fcfaeace-bfb8-457f-8817-c4404d51e37d" (UID: "fcfaeace-bfb8-457f-8817-c4404d51e37d"). InnerVolumeSpecName "kube-api-access-bkf8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.887431 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fcfaeace-bfb8-457f-8817-c4404d51e37d" (UID: "fcfaeace-bfb8-457f-8817-c4404d51e37d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.942091 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4d3115-9287-41a4-9435-1d34590a5178" path="/var/lib/kubelet/pods/2b4d3115-9287-41a4-9435-1d34590a5178/volumes" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.942732 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c4448a-914d-4e2c-9ad7-71a2d0444669" path="/var/lib/kubelet/pods/c8c4448a-914d-4e2c-9ad7-71a2d0444669/volumes" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.985548 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcfaeace-bfb8-457f-8817-c4404d51e37d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.985585 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkf8k\" (UniqueName: \"kubernetes.io/projected/fcfaeace-bfb8-457f-8817-c4404d51e37d-kube-api-access-bkf8k\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.985597 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5j4x\" (UniqueName: \"kubernetes.io/projected/61b0f49c-ce6e-4dd2-be73-75451006164f-kube-api-access-r5j4x\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:49 crc kubenswrapper[4989]: I1006 08:51:49.985605 4989 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61b0f49c-ce6e-4dd2-be73-75451006164f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.854572 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cf875bb9f-wsn4m" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.854572 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.895231 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.896172 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900367 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900491 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900381 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900822 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900834 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.900901 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.901026 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.924491 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-677bf5f74d-2bvcf"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.927459 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.934587 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cf875bb9f-wsn4m"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.936566 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-cf875bb9f-wsn4m"] Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.996447 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-config\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.996498 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-client-ca\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.996535 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkctb\" (UniqueName: \"kubernetes.io/projected/9f790769-f7eb-49da-af1f-9e1ad5edc961-kube-api-access-nkctb\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:50 crc kubenswrapper[4989]: I1006 08:51:50.996576 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f790769-f7eb-49da-af1f-9e1ad5edc961-serving-cert\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.098094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-config\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.098153 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-client-ca\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.098219 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkctb\" (UniqueName: \"kubernetes.io/projected/9f790769-f7eb-49da-af1f-9e1ad5edc961-kube-api-access-nkctb\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.098706 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f790769-f7eb-49da-af1f-9e1ad5edc961-serving-cert\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.099133 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-client-ca\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.099301 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f790769-f7eb-49da-af1f-9e1ad5edc961-config\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.101850 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f790769-f7eb-49da-af1f-9e1ad5edc961-serving-cert\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.114280 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkctb\" (UniqueName: \"kubernetes.io/projected/9f790769-f7eb-49da-af1f-9e1ad5edc961-kube-api-access-nkctb\") pod \"route-controller-manager-69588767fb-xw4dm\" (UID: \"9f790769-f7eb-49da-af1f-9e1ad5edc961\") " pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.212609 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.610456 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm"] Oct 06 08:51:51 crc kubenswrapper[4989]: W1006 08:51:51.617293 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f790769_f7eb_49da_af1f_9e1ad5edc961.slice/crio-d2d01852b7da824e88723bbf89b8ae545be44f5103c3eee828d349e9ded2b22b WatchSource:0}: Error finding container d2d01852b7da824e88723bbf89b8ae545be44f5103c3eee828d349e9ded2b22b: Status 404 returned error can't find the container with id d2d01852b7da824e88723bbf89b8ae545be44f5103c3eee828d349e9ded2b22b Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.860439 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" event={"ID":"9f790769-f7eb-49da-af1f-9e1ad5edc961","Type":"ContainerStarted","Data":"6e03296adb0af43dad3d6a5e8e87d03ad0108875e3547a88d8954e1ee7d63fc4"} Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.860485 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" event={"ID":"9f790769-f7eb-49da-af1f-9e1ad5edc961","Type":"ContainerStarted","Data":"d2d01852b7da824e88723bbf89b8ae545be44f5103c3eee828d349e9ded2b22b"} Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.860720 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.878508 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" podStartSLOduration=2.878485795 podStartE2EDuration="2.878485795s" podCreationTimestamp="2025-10-06 08:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:51:51.876269073 +0000 UTC m=+762.666294683" watchObservedRunningTime="2025-10-06 08:51:51.878485795 +0000 UTC m=+762.668511396" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.941721 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61b0f49c-ce6e-4dd2-be73-75451006164f" path="/var/lib/kubelet/pods/61b0f49c-ce6e-4dd2-be73-75451006164f/volumes" Oct 06 08:51:51 crc kubenswrapper[4989]: I1006 08:51:51.942072 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcfaeace-bfb8-457f-8817-c4404d51e37d" path="/var/lib/kubelet/pods/fcfaeace-bfb8-457f-8817-c4404d51e37d/volumes" Oct 06 08:51:52 crc kubenswrapper[4989]: I1006 08:51:52.399264 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69588767fb-xw4dm" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.019393 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8489bf87df-9tnpv"] Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.020629 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.022372 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.023467 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.027041 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.027044 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.027044 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.027107 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.033469 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.034128 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8489bf87df-9tnpv"] Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.124423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac2f414-88d5-417d-bc6c-758d287aa9c8-serving-cert\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.124496 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-proxy-ca-bundles\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.124558 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjh2q\" (UniqueName: \"kubernetes.io/projected/0ac2f414-88d5-417d-bc6c-758d287aa9c8-kube-api-access-kjh2q\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.124616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-client-ca\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.124644 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-config\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.225968 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjh2q\" (UniqueName: \"kubernetes.io/projected/0ac2f414-88d5-417d-bc6c-758d287aa9c8-kube-api-access-kjh2q\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.226051 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-client-ca\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.226079 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-config\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.226120 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac2f414-88d5-417d-bc6c-758d287aa9c8-serving-cert\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.226161 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-proxy-ca-bundles\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.226867 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-client-ca\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.227255 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-proxy-ca-bundles\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.229439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ac2f414-88d5-417d-bc6c-758d287aa9c8-config\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.234842 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac2f414-88d5-417d-bc6c-758d287aa9c8-serving-cert\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.248863 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjh2q\" (UniqueName: \"kubernetes.io/projected/0ac2f414-88d5-417d-bc6c-758d287aa9c8-kube-api-access-kjh2q\") pod \"controller-manager-8489bf87df-9tnpv\" (UID: \"0ac2f414-88d5-417d-bc6c-758d287aa9c8\") " pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.344208 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.596127 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8489bf87df-9tnpv"] Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.876166 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" event={"ID":"0ac2f414-88d5-417d-bc6c-758d287aa9c8","Type":"ContainerStarted","Data":"d82b1d36958a17b3187608e906de5bfc84def4f1c71036f190abfef12d258af0"} Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.876254 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" event={"ID":"0ac2f414-88d5-417d-bc6c-758d287aa9c8","Type":"ContainerStarted","Data":"e8766788529bf8a4ae45bc5525b24bc3124b2db6ce352c85cbd5ac16bbf9e13b"} Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.907205 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" podStartSLOduration=4.907184387 podStartE2EDuration="4.907184387s" podCreationTimestamp="2025-10-06 08:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:51:53.904951735 +0000 UTC m=+764.694977335" watchObservedRunningTime="2025-10-06 08:51:53.907184387 +0000 UTC m=+764.697209967" Oct 06 08:51:53 crc kubenswrapper[4989]: I1006 08:51:53.981419 4989 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 06 08:51:54 crc kubenswrapper[4989]: I1006 08:51:54.880643 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:51:54 crc kubenswrapper[4989]: I1006 08:51:54.884922 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8489bf87df-9tnpv" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.033855 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.035691 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.047448 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.086342 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.086578 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdjhg\" (UniqueName: \"kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.086684 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.187619 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdjhg\" (UniqueName: \"kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.187822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.187994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.188317 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.188807 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.210875 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdjhg\" (UniqueName: \"kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg\") pod \"certified-operators-sjqtm\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.358887 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.808030 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.950748 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerStarted","Data":"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0"} Oct 06 08:52:05 crc kubenswrapper[4989]: I1006 08:52:05.950955 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerStarted","Data":"e3f5031677981ca9c55f01a509e1f92cb3e0087ea54dafcc5a6e96941a284a8d"} Oct 06 08:52:06 crc kubenswrapper[4989]: I1006 08:52:06.961864 4989 generic.go:334] "Generic (PLEG): container finished" podID="565b363f-f96b-4c05-bab7-434c09790baa" containerID="d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0" exitCode=0 Oct 06 08:52:06 crc kubenswrapper[4989]: I1006 08:52:06.961940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerDied","Data":"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0"} Oct 06 08:52:06 crc kubenswrapper[4989]: I1006 08:52:06.964811 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 08:52:07 crc kubenswrapper[4989]: I1006 08:52:07.968071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerStarted","Data":"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c"} Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.622258 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.625666 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.631357 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.730551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whp6\" (UniqueName: \"kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.730593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.730688 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.832265 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.832423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.832481 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8whp6\" (UniqueName: \"kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.833041 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.833156 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.857138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whp6\" (UniqueName: \"kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6\") pod \"redhat-marketplace-hstpl\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.948853 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.975197 4989 generic.go:334] "Generic (PLEG): container finished" podID="565b363f-f96b-4c05-bab7-434c09790baa" containerID="05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c" exitCode=0 Oct 06 08:52:08 crc kubenswrapper[4989]: I1006 08:52:08.975255 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerDied","Data":"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c"} Oct 06 08:52:09 crc kubenswrapper[4989]: I1006 08:52:09.366110 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:09 crc kubenswrapper[4989]: W1006 08:52:09.374075 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8eb9b26_0feb_4fcf_9dcf_3d4e7f672a0f.slice/crio-e52e7acb883f4439c3349c8a1f5bb72b738f05f83a2b6cd2d528f1078069bd66 WatchSource:0}: Error finding container e52e7acb883f4439c3349c8a1f5bb72b738f05f83a2b6cd2d528f1078069bd66: Status 404 returned error can't find the container with id e52e7acb883f4439c3349c8a1f5bb72b738f05f83a2b6cd2d528f1078069bd66 Oct 06 08:52:09 crc kubenswrapper[4989]: I1006 08:52:09.980486 4989 generic.go:334] "Generic (PLEG): container finished" podID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerID="b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7" exitCode=0 Oct 06 08:52:09 crc kubenswrapper[4989]: I1006 08:52:09.980543 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerDied","Data":"b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7"} Oct 06 08:52:09 crc kubenswrapper[4989]: I1006 08:52:09.980884 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerStarted","Data":"e52e7acb883f4439c3349c8a1f5bb72b738f05f83a2b6cd2d528f1078069bd66"} Oct 06 08:52:09 crc kubenswrapper[4989]: I1006 08:52:09.983103 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerStarted","Data":"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb"} Oct 06 08:52:10 crc kubenswrapper[4989]: I1006 08:52:10.022941 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sjqtm" podStartSLOduration=2.345169915 podStartE2EDuration="5.022922907s" podCreationTimestamp="2025-10-06 08:52:05 +0000 UTC" firstStartedPulling="2025-10-06 08:52:06.964389863 +0000 UTC m=+777.754415483" lastFinishedPulling="2025-10-06 08:52:09.642142885 +0000 UTC m=+780.432168475" observedRunningTime="2025-10-06 08:52:10.019743257 +0000 UTC m=+780.809768827" watchObservedRunningTime="2025-10-06 08:52:10.022922907 +0000 UTC m=+780.812948487" Oct 06 08:52:10 crc kubenswrapper[4989]: I1006 08:52:10.990816 4989 generic.go:334] "Generic (PLEG): container finished" podID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerID="54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c" exitCode=0 Oct 06 08:52:10 crc kubenswrapper[4989]: I1006 08:52:10.992262 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerDied","Data":"54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c"} Oct 06 08:52:13 crc kubenswrapper[4989]: I1006 08:52:13.010329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerStarted","Data":"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14"} Oct 06 08:52:13 crc kubenswrapper[4989]: I1006 08:52:13.035255 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hstpl" podStartSLOduration=3.13442885 podStartE2EDuration="5.035218865s" podCreationTimestamp="2025-10-06 08:52:08 +0000 UTC" firstStartedPulling="2025-10-06 08:52:09.981722028 +0000 UTC m=+780.771747608" lastFinishedPulling="2025-10-06 08:52:11.882512043 +0000 UTC m=+782.672537623" observedRunningTime="2025-10-06 08:52:13.030894843 +0000 UTC m=+783.820920443" watchObservedRunningTime="2025-10-06 08:52:13.035218865 +0000 UTC m=+783.825244515" Oct 06 08:52:15 crc kubenswrapper[4989]: I1006 08:52:15.360069 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:15 crc kubenswrapper[4989]: I1006 08:52:15.361312 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:15 crc kubenswrapper[4989]: I1006 08:52:15.407034 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:16 crc kubenswrapper[4989]: I1006 08:52:16.101859 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:16 crc kubenswrapper[4989]: I1006 08:52:16.622232 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.049473 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sjqtm" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="registry-server" containerID="cri-o://c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb" gracePeriod=2 Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.529721 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.565002 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content\") pod \"565b363f-f96b-4c05-bab7-434c09790baa\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.565224 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities\") pod \"565b363f-f96b-4c05-bab7-434c09790baa\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.566336 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdjhg\" (UniqueName: \"kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg\") pod \"565b363f-f96b-4c05-bab7-434c09790baa\" (UID: \"565b363f-f96b-4c05-bab7-434c09790baa\") " Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.568275 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities" (OuterVolumeSpecName: "utilities") pod "565b363f-f96b-4c05-bab7-434c09790baa" (UID: "565b363f-f96b-4c05-bab7-434c09790baa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.575892 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg" (OuterVolumeSpecName: "kube-api-access-gdjhg") pod "565b363f-f96b-4c05-bab7-434c09790baa" (UID: "565b363f-f96b-4c05-bab7-434c09790baa"). InnerVolumeSpecName "kube-api-access-gdjhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.662811 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "565b363f-f96b-4c05-bab7-434c09790baa" (UID: "565b363f-f96b-4c05-bab7-434c09790baa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.668408 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.668471 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdjhg\" (UniqueName: \"kubernetes.io/projected/565b363f-f96b-4c05-bab7-434c09790baa-kube-api-access-gdjhg\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.668490 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/565b363f-f96b-4c05-bab7-434c09790baa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.949476 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:18 crc kubenswrapper[4989]: I1006 08:52:18.949568 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.012770 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.056963 4989 generic.go:334] "Generic (PLEG): container finished" podID="565b363f-f96b-4c05-bab7-434c09790baa" containerID="c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb" exitCode=0 Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.057087 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjqtm" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.057148 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerDied","Data":"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb"} Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.057185 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjqtm" event={"ID":"565b363f-f96b-4c05-bab7-434c09790baa","Type":"ContainerDied","Data":"e3f5031677981ca9c55f01a509e1f92cb3e0087ea54dafcc5a6e96941a284a8d"} Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.057205 4989 scope.go:117] "RemoveContainer" containerID="c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.077773 4989 scope.go:117] "RemoveContainer" containerID="05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.093416 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.093468 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sjqtm"] Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.120042 4989 scope.go:117] "RemoveContainer" containerID="d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.122010 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.145626 4989 scope.go:117] "RemoveContainer" containerID="c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb" Oct 06 08:52:19 crc kubenswrapper[4989]: E1006 08:52:19.146146 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb\": container with ID starting with c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb not found: ID does not exist" containerID="c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.146189 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb"} err="failed to get container status \"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb\": rpc error: code = NotFound desc = could not find container \"c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb\": container with ID starting with c5de5cb198e94c07a9ef90fbc26379121d78d0efafc91d8f10c3a5514ffa16cb not found: ID does not exist" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.146230 4989 scope.go:117] "RemoveContainer" containerID="05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c" Oct 06 08:52:19 crc kubenswrapper[4989]: E1006 08:52:19.146693 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c\": container with ID starting with 05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c not found: ID does not exist" containerID="05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.146759 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c"} err="failed to get container status \"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c\": rpc error: code = NotFound desc = could not find container \"05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c\": container with ID starting with 05fafd08d59e2939549c18fef65b0c9e0e71b04c08a9a89ff8885d09de74148c not found: ID does not exist" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.146787 4989 scope.go:117] "RemoveContainer" containerID="d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0" Oct 06 08:52:19 crc kubenswrapper[4989]: E1006 08:52:19.147078 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0\": container with ID starting with d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0 not found: ID does not exist" containerID="d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.147130 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0"} err="failed to get container status \"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0\": rpc error: code = NotFound desc = could not find container \"d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0\": container with ID starting with d557b8867fc4da61ec4c81c0380a62d2a6eaf1fb2296624ee33b97fdcf6d61d0 not found: ID does not exist" Oct 06 08:52:19 crc kubenswrapper[4989]: I1006 08:52:19.943940 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565b363f-f96b-4c05-bab7-434c09790baa" path="/var/lib/kubelet/pods/565b363f-f96b-4c05-bab7-434c09790baa/volumes" Oct 06 08:52:21 crc kubenswrapper[4989]: I1006 08:52:21.413852 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:21 crc kubenswrapper[4989]: I1006 08:52:21.414069 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hstpl" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="registry-server" containerID="cri-o://d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14" gracePeriod=2 Oct 06 08:52:21 crc kubenswrapper[4989]: I1006 08:52:21.889147 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.020927 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities\") pod \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.021005 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8whp6\" (UniqueName: \"kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6\") pod \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.021116 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content\") pod \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\" (UID: \"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f\") " Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.021847 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities" (OuterVolumeSpecName: "utilities") pod "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" (UID: "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.027235 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6" (OuterVolumeSpecName: "kube-api-access-8whp6") pod "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" (UID: "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f"). InnerVolumeSpecName "kube-api-access-8whp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.032818 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" (UID: "e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.077133 4989 generic.go:334] "Generic (PLEG): container finished" podID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerID="d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14" exitCode=0 Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.077198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerDied","Data":"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14"} Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.077222 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hstpl" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.077246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hstpl" event={"ID":"e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f","Type":"ContainerDied","Data":"e52e7acb883f4439c3349c8a1f5bb72b738f05f83a2b6cd2d528f1078069bd66"} Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.077267 4989 scope.go:117] "RemoveContainer" containerID="d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.096407 4989 scope.go:117] "RemoveContainer" containerID="54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.103251 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.111190 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hstpl"] Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.122549 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.122770 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8whp6\" (UniqueName: \"kubernetes.io/projected/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-kube-api-access-8whp6\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.122859 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.133952 4989 scope.go:117] "RemoveContainer" containerID="b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.151188 4989 scope.go:117] "RemoveContainer" containerID="d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14" Oct 06 08:52:22 crc kubenswrapper[4989]: E1006 08:52:22.151854 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14\": container with ID starting with d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14 not found: ID does not exist" containerID="d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.151893 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14"} err="failed to get container status \"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14\": rpc error: code = NotFound desc = could not find container \"d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14\": container with ID starting with d7eeab3127cd99fef56fe6e96b14c163c278b23f1c67e5a592a089c89fa9ec14 not found: ID does not exist" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.151921 4989 scope.go:117] "RemoveContainer" containerID="54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c" Oct 06 08:52:22 crc kubenswrapper[4989]: E1006 08:52:22.152393 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c\": container with ID starting with 54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c not found: ID does not exist" containerID="54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.152501 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c"} err="failed to get container status \"54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c\": rpc error: code = NotFound desc = could not find container \"54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c\": container with ID starting with 54ff73abe25e95021f79d2b78cf645e6fb8e2f2e64ce43bc6917b09e8e75a25c not found: ID does not exist" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.152592 4989 scope.go:117] "RemoveContainer" containerID="b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7" Oct 06 08:52:22 crc kubenswrapper[4989]: E1006 08:52:22.153217 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7\": container with ID starting with b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7 not found: ID does not exist" containerID="b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7" Oct 06 08:52:22 crc kubenswrapper[4989]: I1006 08:52:22.153258 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7"} err="failed to get container status \"b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7\": rpc error: code = NotFound desc = could not find container \"b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7\": container with ID starting with b6e46dda105d9cf1101d4bb9aba1b007231c10a7337e0bed22d1b4582e37a6d7 not found: ID does not exist" Oct 06 08:52:23 crc kubenswrapper[4989]: I1006 08:52:23.948575 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" path="/var/lib/kubelet/pods/e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f/volumes" Oct 06 08:52:33 crc kubenswrapper[4989]: I1006 08:52:33.935689 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:52:33 crc kubenswrapper[4989]: I1006 08:52:33.936440 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:53:03 crc kubenswrapper[4989]: I1006 08:53:03.935264 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:53:03 crc kubenswrapper[4989]: I1006 08:53:03.935980 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:53:33 crc kubenswrapper[4989]: I1006 08:53:33.936088 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:53:33 crc kubenswrapper[4989]: I1006 08:53:33.936630 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:53:33 crc kubenswrapper[4989]: I1006 08:53:33.942253 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:53:33 crc kubenswrapper[4989]: I1006 08:53:33.942878 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 08:53:33 crc kubenswrapper[4989]: I1006 08:53:33.942968 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a" gracePeriod=600 Oct 06 08:53:34 crc kubenswrapper[4989]: I1006 08:53:34.514460 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a" exitCode=0 Oct 06 08:53:34 crc kubenswrapper[4989]: I1006 08:53:34.514486 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a"} Oct 06 08:53:34 crc kubenswrapper[4989]: I1006 08:53:34.514919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f"} Oct 06 08:53:34 crc kubenswrapper[4989]: I1006 08:53:34.514948 4989 scope.go:117] "RemoveContainer" containerID="696761c7cfa8ae1fd9d3500f712f647850e1391d7ed42e6f8c628155604422e2" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.554862 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t6mdb"] Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555799 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="nbdb" containerID="cri-o://636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555820 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555790 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-controller" containerID="cri-o://9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555906 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-acl-logging" containerID="cri-o://cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555899 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="northd" containerID="cri-o://52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.555936 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="sbdb" containerID="cri-o://10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.556020 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-node" containerID="cri-o://130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.584757 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" containerID="cri-o://afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" gracePeriod=30 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.674519 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/2.log" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.675098 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/1.log" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.675176 4989 generic.go:334] "Generic (PLEG): container finished" podID="b901f1ca-c342-4faa-a715-0ff3d94bc369" containerID="6cd6e8e428a6d15f8ae746ce43a6eaa2ff4c01b5157090e193f58fe5ac9cdb48" exitCode=2 Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.675233 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerDied","Data":"6cd6e8e428a6d15f8ae746ce43a6eaa2ff4c01b5157090e193f58fe5ac9cdb48"} Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.675289 4989 scope.go:117] "RemoveContainer" containerID="2ed98c0a531b5452e2bd62efe11c4dcdf9dbafe1028760a80d86cdc4718de7aa" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.675841 4989 scope.go:117] "RemoveContainer" containerID="6cd6e8e428a6d15f8ae746ce43a6eaa2ff4c01b5157090e193f58fe5ac9cdb48" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.891374 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/3.log" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.895096 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovn-acl-logging/0.log" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.896644 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovn-controller/0.log" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.897089 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.953890 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4699c"] Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954180 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="nbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954197 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="nbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954208 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="northd" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954214 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="northd" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954221 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="extract-content" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954229 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="extract-content" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954243 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="sbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954249 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="sbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954259 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954266 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954277 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954284 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954292 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="extract-utilities" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954299 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="extract-utilities" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954307 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="extract-content" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954313 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="extract-content" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954321 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954327 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954335 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954341 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954349 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-acl-logging" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954355 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-acl-logging" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954362 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kubecfg-setup" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954369 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kubecfg-setup" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954376 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954381 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954387 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-node" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954394 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-node" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954400 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954406 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954415 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954422 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954434 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954439 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954448 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="extract-utilities" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954455 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="extract-utilities" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954560 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8eb9b26-0feb-4fcf-9dcf-3d4e7f672a0f" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954568 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954576 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954584 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="sbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954592 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954599 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954610 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954620 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-ovn-metrics" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954627 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="northd" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954636 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="565b363f-f96b-4c05-bab7-434c09790baa" containerName="registry-server" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954644 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="kube-rbac-proxy-node" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954724 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovn-acl-logging" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954734 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="nbdb" Oct 06 08:53:57 crc kubenswrapper[4989]: E1006 08:53:57.954818 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954825 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.954948 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerName="ovnkube-controller" Oct 06 08:53:57 crc kubenswrapper[4989]: I1006 08:53:57.956685 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.070900 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.070956 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.070992 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071033 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071054 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071075 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071066 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071091 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071171 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071201 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071226 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071274 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071291 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071330 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071349 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071375 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071405 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071423 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071446 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nl6q\" (UniqueName: \"kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071472 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet\") pod \"91d0cca1-dfc6-4f12-a25a-7325782935ba\" (UID: \"91d0cca1-dfc6-4f12-a25a-7325782935ba\") " Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-config\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071704 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-node-log\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071764 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-systemd-units\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071794 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-kubelet\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071817 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-ovn\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071855 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071880 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-netns\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-netd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071957 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-etc-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072009 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-script-lib\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072070 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-env-overrides\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072092 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9ttx\" (UniqueName: \"kubernetes.io/projected/d952eb09-1330-451b-8a7f-61cb3eb0d113-kube-api-access-h9ttx\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072122 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovn-node-metrics-cert\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072141 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-var-lib-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072169 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-bin\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072206 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-log-socket\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-slash\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-systemd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072277 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072319 4989 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072333 4989 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071172 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072711 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071539 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071557 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071569 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log" (OuterVolumeSpecName: "node-log") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071582 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.071833 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072043 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072384 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072403 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket" (OuterVolumeSpecName: "log-socket") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072583 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072776 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072690 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.072709 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.073142 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash" (OuterVolumeSpecName: "host-slash") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.078266 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q" (OuterVolumeSpecName: "kube-api-access-6nl6q") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "kube-api-access-6nl6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.078613 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.086879 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "91d0cca1-dfc6-4f12-a25a-7325782935ba" (UID: "91d0cca1-dfc6-4f12-a25a-7325782935ba"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173194 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovn-node-metrics-cert\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173237 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-var-lib-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173260 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-bin\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173285 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-log-socket\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173303 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-slash\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173318 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-systemd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173349 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-config\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173348 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-var-lib-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173406 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-bin\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173414 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-node-log\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173368 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-node-log\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173441 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-log-socket\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173464 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-slash\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173468 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173483 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-systemd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173502 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173502 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-systemd-units\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173527 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-systemd-units\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173532 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-kubelet\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-kubelet\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173563 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-ovn\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173613 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173640 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-netns\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173688 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-netd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173705 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-run-ovn\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173718 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-etc-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173749 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173781 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-etc-openvswitch\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173794 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-script-lib\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173814 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-ovn-kubernetes\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173840 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-run-netns\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-env-overrides\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173869 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d952eb09-1330-451b-8a7f-61cb3eb0d113-host-cni-netd\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173879 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9ttx\" (UniqueName: \"kubernetes.io/projected/d952eb09-1330-451b-8a7f-61cb3eb0d113-kube-api-access-h9ttx\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173951 4989 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173964 4989 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173976 4989 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173987 4989 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.173996 4989 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174004 4989 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174012 4989 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-log-socket\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174022 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nl6q\" (UniqueName: \"kubernetes.io/projected/91d0cca1-dfc6-4f12-a25a-7325782935ba-kube-api-access-6nl6q\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174032 4989 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174052 4989 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174060 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174070 4989 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174078 4989 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-node-log\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174085 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174093 4989 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174101 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/91d0cca1-dfc6-4f12-a25a-7325782935ba-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174108 4989 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174117 4989 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/91d0cca1-dfc6-4f12-a25a-7325782935ba-host-slash\") on node \"crc\" DevicePath \"\"" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-config\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174467 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovnkube-script-lib\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.174560 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d952eb09-1330-451b-8a7f-61cb3eb0d113-env-overrides\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.177170 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d952eb09-1330-451b-8a7f-61cb3eb0d113-ovn-node-metrics-cert\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.195606 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9ttx\" (UniqueName: \"kubernetes.io/projected/d952eb09-1330-451b-8a7f-61cb3eb0d113-kube-api-access-h9ttx\") pod \"ovnkube-node-4699c\" (UID: \"d952eb09-1330-451b-8a7f-61cb3eb0d113\") " pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.270957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:53:58 crc kubenswrapper[4989]: W1006 08:53:58.288160 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd952eb09_1330_451b_8a7f_61cb3eb0d113.slice/crio-516ca8ba4a6c84a082b9dd3a31885aa21578a99d53a8ca948c8beafe732c10cb WatchSource:0}: Error finding container 516ca8ba4a6c84a082b9dd3a31885aa21578a99d53a8ca948c8beafe732c10cb: Status 404 returned error can't find the container with id 516ca8ba4a6c84a082b9dd3a31885aa21578a99d53a8ca948c8beafe732c10cb Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.683150 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wz8dx_b901f1ca-c342-4faa-a715-0ff3d94bc369/kube-multus/2.log" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.683228 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wz8dx" event={"ID":"b901f1ca-c342-4faa-a715-0ff3d94bc369","Type":"ContainerStarted","Data":"a034533d2b5bc7fabdc97e4008e62a6bd85f6316b3bc0b6c4326c572287380bc"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.687569 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovnkube-controller/3.log" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.690101 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovn-acl-logging/0.log" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.690575 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-t6mdb_91d0cca1-dfc6-4f12-a25a-7325782935ba/ovn-controller/0.log" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691397 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691420 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691430 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691437 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691443 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691450 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691456 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" exitCode=143 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691462 4989 generic.go:334] "Generic (PLEG): container finished" podID="91d0cca1-dfc6-4f12-a25a-7325782935ba" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" exitCode=143 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691505 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691521 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691559 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691569 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691580 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691588 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691598 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691608 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691613 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691619 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691663 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691674 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691681 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691687 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691693 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691702 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691737 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691748 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691754 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691759 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691764 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691772 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691776 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691781 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691786 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691791 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691819 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691829 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691835 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691840 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691845 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691849 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691854 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691859 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691864 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691869 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691874 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691903 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-t6mdb" event={"ID":"91d0cca1-dfc6-4f12-a25a-7325782935ba","Type":"ContainerDied","Data":"759d4d68f08c2d705bd6ef7f317cb1783cf09419ac82808310903662d1285dcd"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691911 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691917 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691922 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691928 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691934 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691939 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691944 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691949 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691953 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691978 4989 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.691992 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.693756 4989 generic.go:334] "Generic (PLEG): container finished" podID="d952eb09-1330-451b-8a7f-61cb3eb0d113" containerID="de1a77a5859d7c866b453d6ef1b8d0dc083788ed4662ca1836978eaf8a4cfba4" exitCode=0 Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.693778 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerDied","Data":"de1a77a5859d7c866b453d6ef1b8d0dc083788ed4662ca1836978eaf8a4cfba4"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.693795 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"516ca8ba4a6c84a082b9dd3a31885aa21578a99d53a8ca948c8beafe732c10cb"} Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.716774 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.769347 4989 scope.go:117] "RemoveContainer" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.773722 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t6mdb"] Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.785344 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-t6mdb"] Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.808852 4989 scope.go:117] "RemoveContainer" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.840683 4989 scope.go:117] "RemoveContainer" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.856162 4989 scope.go:117] "RemoveContainer" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.870927 4989 scope.go:117] "RemoveContainer" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.887682 4989 scope.go:117] "RemoveContainer" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.913214 4989 scope.go:117] "RemoveContainer" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.933334 4989 scope.go:117] "RemoveContainer" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.960295 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.961031 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961067 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} err="failed to get container status \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961094 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.961431 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": container with ID starting with ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53 not found: ID does not exist" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961454 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} err="failed to get container status \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": rpc error: code = NotFound desc = could not find container \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": container with ID starting with ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961469 4989 scope.go:117] "RemoveContainer" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.961799 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": container with ID starting with 10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25 not found: ID does not exist" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961824 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} err="failed to get container status \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": rpc error: code = NotFound desc = could not find container \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": container with ID starting with 10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.961844 4989 scope.go:117] "RemoveContainer" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.962239 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": container with ID starting with 636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e not found: ID does not exist" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.962296 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} err="failed to get container status \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": rpc error: code = NotFound desc = could not find container \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": container with ID starting with 636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.962330 4989 scope.go:117] "RemoveContainer" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.962858 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": container with ID starting with 52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770 not found: ID does not exist" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.962888 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} err="failed to get container status \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": rpc error: code = NotFound desc = could not find container \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": container with ID starting with 52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.962908 4989 scope.go:117] "RemoveContainer" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.963325 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": container with ID starting with 61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f not found: ID does not exist" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.963358 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} err="failed to get container status \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": rpc error: code = NotFound desc = could not find container \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": container with ID starting with 61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.963382 4989 scope.go:117] "RemoveContainer" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.963667 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": container with ID starting with 130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553 not found: ID does not exist" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.963696 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} err="failed to get container status \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": rpc error: code = NotFound desc = could not find container \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": container with ID starting with 130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.963715 4989 scope.go:117] "RemoveContainer" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.963997 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": container with ID starting with cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a not found: ID does not exist" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.964031 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} err="failed to get container status \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": rpc error: code = NotFound desc = could not find container \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": container with ID starting with cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.964052 4989 scope.go:117] "RemoveContainer" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.964790 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": container with ID starting with 9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f not found: ID does not exist" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.964818 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} err="failed to get container status \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": rpc error: code = NotFound desc = could not find container \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": container with ID starting with 9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.964864 4989 scope.go:117] "RemoveContainer" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: E1006 08:53:58.965321 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": container with ID starting with 077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478 not found: ID does not exist" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.965347 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} err="failed to get container status \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": rpc error: code = NotFound desc = could not find container \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": container with ID starting with 077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.965364 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.965691 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} err="failed to get container status \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.965710 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.965982 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} err="failed to get container status \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": rpc error: code = NotFound desc = could not find container \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": container with ID starting with ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.966011 4989 scope.go:117] "RemoveContainer" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.966326 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} err="failed to get container status \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": rpc error: code = NotFound desc = could not find container \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": container with ID starting with 10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.966350 4989 scope.go:117] "RemoveContainer" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.966587 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} err="failed to get container status \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": rpc error: code = NotFound desc = could not find container \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": container with ID starting with 636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.966610 4989 scope.go:117] "RemoveContainer" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967010 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} err="failed to get container status \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": rpc error: code = NotFound desc = could not find container \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": container with ID starting with 52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967034 4989 scope.go:117] "RemoveContainer" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967260 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} err="failed to get container status \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": rpc error: code = NotFound desc = could not find container \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": container with ID starting with 61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967287 4989 scope.go:117] "RemoveContainer" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967538 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} err="failed to get container status \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": rpc error: code = NotFound desc = could not find container \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": container with ID starting with 130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967561 4989 scope.go:117] "RemoveContainer" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967829 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} err="failed to get container status \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": rpc error: code = NotFound desc = could not find container \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": container with ID starting with cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.967853 4989 scope.go:117] "RemoveContainer" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.969274 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} err="failed to get container status \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": rpc error: code = NotFound desc = could not find container \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": container with ID starting with 9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.969300 4989 scope.go:117] "RemoveContainer" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.969643 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} err="failed to get container status \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": rpc error: code = NotFound desc = could not find container \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": container with ID starting with 077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.969685 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970039 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} err="failed to get container status \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970057 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970347 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} err="failed to get container status \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": rpc error: code = NotFound desc = could not find container \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": container with ID starting with ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970378 4989 scope.go:117] "RemoveContainer" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970693 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} err="failed to get container status \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": rpc error: code = NotFound desc = could not find container \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": container with ID starting with 10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.970735 4989 scope.go:117] "RemoveContainer" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972069 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} err="failed to get container status \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": rpc error: code = NotFound desc = could not find container \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": container with ID starting with 636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972108 4989 scope.go:117] "RemoveContainer" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972566 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} err="failed to get container status \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": rpc error: code = NotFound desc = could not find container \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": container with ID starting with 52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972589 4989 scope.go:117] "RemoveContainer" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972827 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} err="failed to get container status \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": rpc error: code = NotFound desc = could not find container \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": container with ID starting with 61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.972846 4989 scope.go:117] "RemoveContainer" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973075 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} err="failed to get container status \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": rpc error: code = NotFound desc = could not find container \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": container with ID starting with 130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973097 4989 scope.go:117] "RemoveContainer" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973320 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} err="failed to get container status \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": rpc error: code = NotFound desc = could not find container \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": container with ID starting with cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973341 4989 scope.go:117] "RemoveContainer" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973699 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} err="failed to get container status \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": rpc error: code = NotFound desc = could not find container \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": container with ID starting with 9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.973726 4989 scope.go:117] "RemoveContainer" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.974808 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} err="failed to get container status \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": rpc error: code = NotFound desc = could not find container \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": container with ID starting with 077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.974835 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.975186 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} err="failed to get container status \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.975209 4989 scope.go:117] "RemoveContainer" containerID="ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.975572 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53"} err="failed to get container status \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": rpc error: code = NotFound desc = could not find container \"ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53\": container with ID starting with ddf7cf50ef7f7b10d73808f53ff9608028ce28744997dc0f12e138b10e568d53 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.975596 4989 scope.go:117] "RemoveContainer" containerID="10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.975998 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25"} err="failed to get container status \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": rpc error: code = NotFound desc = could not find container \"10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25\": container with ID starting with 10f6b62d1f2538ee0f19910a3fdca90112ecc793f6621bcc16d300248620dd25 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976019 4989 scope.go:117] "RemoveContainer" containerID="636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976332 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e"} err="failed to get container status \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": rpc error: code = NotFound desc = could not find container \"636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e\": container with ID starting with 636d1a8d72881f8f17ffa3a0107bab68b9fbc90cc6fa70e4d412a5c7a6f4027e not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976354 4989 scope.go:117] "RemoveContainer" containerID="52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976600 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770"} err="failed to get container status \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": rpc error: code = NotFound desc = could not find container \"52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770\": container with ID starting with 52aac5c4327c07533ba16f6d870af9441f65dd3e2d1ea27269d310a4ca1a7770 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976616 4989 scope.go:117] "RemoveContainer" containerID="61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976947 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f"} err="failed to get container status \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": rpc error: code = NotFound desc = could not find container \"61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f\": container with ID starting with 61558ed2b7ad4e1689282100b6ae698d38bd59d3d857e2a4a707a8daec58649f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.976963 4989 scope.go:117] "RemoveContainer" containerID="130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977143 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553"} err="failed to get container status \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": rpc error: code = NotFound desc = could not find container \"130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553\": container with ID starting with 130561384d45e3981b3559649b32a6efda557d70120de33a4670ca4299fb7553 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977158 4989 scope.go:117] "RemoveContainer" containerID="cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977329 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a"} err="failed to get container status \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": rpc error: code = NotFound desc = could not find container \"cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a\": container with ID starting with cdcef7810b08dfabdc12d84c9ada71b3b6bbc33a03502b34dd23e2befb9b585a not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977345 4989 scope.go:117] "RemoveContainer" containerID="9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977529 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f"} err="failed to get container status \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": rpc error: code = NotFound desc = could not find container \"9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f\": container with ID starting with 9c7921fe48989177443bf142caa48283edb0ba5311e4f083d146d5b87abdd98f not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977543 4989 scope.go:117] "RemoveContainer" containerID="077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977751 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478"} err="failed to get container status \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": rpc error: code = NotFound desc = could not find container \"077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478\": container with ID starting with 077055de31b363e5f6323126d122641bc7a64174e991d0fa0d7de19822cf4478 not found: ID does not exist" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.977771 4989 scope.go:117] "RemoveContainer" containerID="afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9" Oct 06 08:53:58 crc kubenswrapper[4989]: I1006 08:53:58.978066 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9"} err="failed to get container status \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": rpc error: code = NotFound desc = could not find container \"afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9\": container with ID starting with afbd93cc0d26945534901603da0914683b3f4853b34a688e3d10275c735a28f9 not found: ID does not exist" Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710468 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"8fc0a50706d8867ec073b0092b94fca7091d9e976cee6192448ac0e57dcd9fab"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710816 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"66187b47894226eb527696b2af25a6f2a91e45f836cdc5a7d02c1a73cb278ed9"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710834 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"ad6605f7701a35601de1a1573725b1af97308fde80c722a2e1fe5bca3ecd2b24"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710850 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"0556a8f5da710b2314ac703841316e2ac38309d05e07d966961a571887ae32de"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710861 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"dbab600a4f69ef51bc6eb1c13b03dcf4f6236fe893fd59907ae135d3e57a18d0"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.710873 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"d1dd43dc4afb3ea1f3d55744f730eeb5b7326347f9fd334ce8409e923bb99e03"} Oct 06 08:53:59 crc kubenswrapper[4989]: I1006 08:53:59.944056 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91d0cca1-dfc6-4f12-a25a-7325782935ba" path="/var/lib/kubelet/pods/91d0cca1-dfc6-4f12-a25a-7325782935ba/volumes" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.572439 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-gfvd9"] Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.573471 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.575686 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.575786 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.575847 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.575795 4989 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-t6whk" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.711881 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.711933 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.711970 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mljbk\" (UniqueName: \"kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.812962 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.813018 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.813068 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mljbk\" (UniqueName: \"kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.813429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.814118 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.848444 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mljbk\" (UniqueName: \"kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk\") pod \"crc-storage-crc-gfvd9\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: I1006 08:54:00.891028 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: E1006 08:54:00.935437 4989 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(e272e2233a2d5c73e301de2a83788f79ac7be55150eb47ee8ed8a1e755ccc098): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:54:00 crc kubenswrapper[4989]: E1006 08:54:00.935531 4989 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(e272e2233a2d5c73e301de2a83788f79ac7be55150eb47ee8ed8a1e755ccc098): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: E1006 08:54:00.935556 4989 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(e272e2233a2d5c73e301de2a83788f79ac7be55150eb47ee8ed8a1e755ccc098): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:00 crc kubenswrapper[4989]: E1006 08:54:00.935598 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-gfvd9_crc-storage(0326ab5a-9c27-4d96-bd43-971561279956)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-gfvd9_crc-storage(0326ab5a-9c27-4d96-bd43-971561279956)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(e272e2233a2d5c73e301de2a83788f79ac7be55150eb47ee8ed8a1e755ccc098): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-gfvd9" podUID="0326ab5a-9c27-4d96-bd43-971561279956" Oct 06 08:54:01 crc kubenswrapper[4989]: I1006 08:54:01.728905 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"e07dfa390f39fcb2ee1cf4098876c7adf31aecc9e09be16a9c26cd8d8920ce8a"} Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.453475 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-gfvd9"] Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.454214 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.454601 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:04 crc kubenswrapper[4989]: E1006 08:54:04.480519 4989 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(6cc580d5943cdf29fc575cc762a50c355fb4becc9864dfb76358ba2091c63f49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 06 08:54:04 crc kubenswrapper[4989]: E1006 08:54:04.480581 4989 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(6cc580d5943cdf29fc575cc762a50c355fb4becc9864dfb76358ba2091c63f49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:04 crc kubenswrapper[4989]: E1006 08:54:04.480600 4989 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(6cc580d5943cdf29fc575cc762a50c355fb4becc9864dfb76358ba2091c63f49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:04 crc kubenswrapper[4989]: E1006 08:54:04.480648 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-gfvd9_crc-storage(0326ab5a-9c27-4d96-bd43-971561279956)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-gfvd9_crc-storage(0326ab5a-9c27-4d96-bd43-971561279956)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-gfvd9_crc-storage_0326ab5a-9c27-4d96-bd43-971561279956_0(6cc580d5943cdf29fc575cc762a50c355fb4becc9864dfb76358ba2091c63f49): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-gfvd9" podUID="0326ab5a-9c27-4d96-bd43-971561279956" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.749141 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" event={"ID":"d952eb09-1330-451b-8a7f-61cb3eb0d113","Type":"ContainerStarted","Data":"70070b81c37a07b3a939d98c0da0a8413578708d616abf4696a90ec62a72dede"} Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.749605 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.749663 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.749677 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.781904 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.785027 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" podStartSLOduration=7.785008195 podStartE2EDuration="7.785008195s" podCreationTimestamp="2025-10-06 08:53:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:54:04.782895171 +0000 UTC m=+895.572920761" watchObservedRunningTime="2025-10-06 08:54:04.785008195 +0000 UTC m=+895.575033775" Oct 06 08:54:04 crc kubenswrapper[4989]: I1006 08:54:04.792287 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:16 crc kubenswrapper[4989]: I1006 08:54:16.936113 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:16 crc kubenswrapper[4989]: I1006 08:54:16.937308 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:17 crc kubenswrapper[4989]: I1006 08:54:17.179433 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-gfvd9"] Oct 06 08:54:17 crc kubenswrapper[4989]: W1006 08:54:17.186921 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0326ab5a_9c27_4d96_bd43_971561279956.slice/crio-a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f WatchSource:0}: Error finding container a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f: Status 404 returned error can't find the container with id a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f Oct 06 08:54:17 crc kubenswrapper[4989]: I1006 08:54:17.836921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gfvd9" event={"ID":"0326ab5a-9c27-4d96-bd43-971561279956","Type":"ContainerStarted","Data":"a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f"} Oct 06 08:54:18 crc kubenswrapper[4989]: I1006 08:54:18.844110 4989 generic.go:334] "Generic (PLEG): container finished" podID="0326ab5a-9c27-4d96-bd43-971561279956" containerID="45967ea767aea12bb58304edab0ce67ec5c5ad5dfd188e799d07704e13309789" exitCode=0 Oct 06 08:54:18 crc kubenswrapper[4989]: I1006 08:54:18.844198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gfvd9" event={"ID":"0326ab5a-9c27-4d96-bd43-971561279956","Type":"ContainerDied","Data":"45967ea767aea12bb58304edab0ce67ec5c5ad5dfd188e799d07704e13309789"} Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.057579 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.059502 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mljbk\" (UniqueName: \"kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk\") pod \"0326ab5a-9c27-4d96-bd43-971561279956\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.059556 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage\") pod \"0326ab5a-9c27-4d96-bd43-971561279956\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.059585 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt\") pod \"0326ab5a-9c27-4d96-bd43-971561279956\" (UID: \"0326ab5a-9c27-4d96-bd43-971561279956\") " Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.059696 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "0326ab5a-9c27-4d96-bd43-971561279956" (UID: "0326ab5a-9c27-4d96-bd43-971561279956"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.059821 4989 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0326ab5a-9c27-4d96-bd43-971561279956-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.065910 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk" (OuterVolumeSpecName: "kube-api-access-mljbk") pod "0326ab5a-9c27-4d96-bd43-971561279956" (UID: "0326ab5a-9c27-4d96-bd43-971561279956"). InnerVolumeSpecName "kube-api-access-mljbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.081866 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "0326ab5a-9c27-4d96-bd43-971561279956" (UID: "0326ab5a-9c27-4d96-bd43-971561279956"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.160810 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mljbk\" (UniqueName: \"kubernetes.io/projected/0326ab5a-9c27-4d96-bd43-971561279956-kube-api-access-mljbk\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.160847 4989 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0326ab5a-9c27-4d96-bd43-971561279956-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.858861 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-gfvd9" event={"ID":"0326ab5a-9c27-4d96-bd43-971561279956","Type":"ContainerDied","Data":"a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f"} Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.859469 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9d1edea5687b5dbad94f7cace66fcb5f08fd322187cf5ff28eca7495bc43a7f" Oct 06 08:54:20 crc kubenswrapper[4989]: I1006 08:54:20.858915 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-gfvd9" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.804255 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g"] Oct 06 08:54:27 crc kubenswrapper[4989]: E1006 08:54:27.805002 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0326ab5a-9c27-4d96-bd43-971561279956" containerName="storage" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.805027 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0326ab5a-9c27-4d96-bd43-971561279956" containerName="storage" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.805213 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0326ab5a-9c27-4d96-bd43-971561279956" containerName="storage" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.806522 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.811756 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.815430 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g"] Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.855507 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.855692 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.855761 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4rhm\" (UniqueName: \"kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.956723 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.957530 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.957628 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.957701 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4rhm\" (UniqueName: \"kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.958048 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:27 crc kubenswrapper[4989]: I1006 08:54:27.983896 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4rhm\" (UniqueName: \"kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:28 crc kubenswrapper[4989]: I1006 08:54:28.123498 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:28 crc kubenswrapper[4989]: I1006 08:54:28.302769 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g"] Oct 06 08:54:28 crc kubenswrapper[4989]: I1006 08:54:28.304802 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4699c" Oct 06 08:54:28 crc kubenswrapper[4989]: W1006 08:54:28.311342 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93fa8693_a1bc_4438_8a3c_318c6374f1b4.slice/crio-29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0 WatchSource:0}: Error finding container 29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0: Status 404 returned error can't find the container with id 29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0 Oct 06 08:54:28 crc kubenswrapper[4989]: I1006 08:54:28.909362 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerStarted","Data":"0c4abf7e6024ec59789150dfb029197239f6e77c1705d23384cab01fee01b47e"} Oct 06 08:54:28 crc kubenswrapper[4989]: I1006 08:54:28.910411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerStarted","Data":"29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0"} Oct 06 08:54:29 crc kubenswrapper[4989]: I1006 08:54:29.917408 4989 generic.go:334] "Generic (PLEG): container finished" podID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerID="0c4abf7e6024ec59789150dfb029197239f6e77c1705d23384cab01fee01b47e" exitCode=0 Oct 06 08:54:29 crc kubenswrapper[4989]: I1006 08:54:29.917938 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerDied","Data":"0c4abf7e6024ec59789150dfb029197239f6e77c1705d23384cab01fee01b47e"} Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.144978 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.146220 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.153054 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.191044 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhm66\" (UniqueName: \"kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.191112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.191136 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.291602 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhm66\" (UniqueName: \"kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.291872 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.291973 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.292470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.292521 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.311727 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhm66\" (UniqueName: \"kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66\") pod \"redhat-operators-klw6n\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.461300 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.853550 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:30 crc kubenswrapper[4989]: W1006 08:54:30.864265 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d2ddb93_a5a3_4eca_8b55_23b60be2f788.slice/crio-250614d6e34bbd29277dc604aad606c77d6e9d7b499cebc71afafb1c1da82105 WatchSource:0}: Error finding container 250614d6e34bbd29277dc604aad606c77d6e9d7b499cebc71afafb1c1da82105: Status 404 returned error can't find the container with id 250614d6e34bbd29277dc604aad606c77d6e9d7b499cebc71afafb1c1da82105 Oct 06 08:54:30 crc kubenswrapper[4989]: I1006 08:54:30.927962 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerStarted","Data":"250614d6e34bbd29277dc604aad606c77d6e9d7b499cebc71afafb1c1da82105"} Oct 06 08:54:31 crc kubenswrapper[4989]: I1006 08:54:31.941646 4989 generic.go:334] "Generic (PLEG): container finished" podID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerID="274be2f716c1747f2f34402bb726ce76fa794edc9e4b86c4f082886ba63db1b0" exitCode=0 Oct 06 08:54:31 crc kubenswrapper[4989]: I1006 08:54:31.945755 4989 generic.go:334] "Generic (PLEG): container finished" podID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerID="f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd" exitCode=0 Oct 06 08:54:31 crc kubenswrapper[4989]: I1006 08:54:31.948925 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerDied","Data":"274be2f716c1747f2f34402bb726ce76fa794edc9e4b86c4f082886ba63db1b0"} Oct 06 08:54:31 crc kubenswrapper[4989]: I1006 08:54:31.948996 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerDied","Data":"f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd"} Oct 06 08:54:32 crc kubenswrapper[4989]: I1006 08:54:32.957744 4989 generic.go:334] "Generic (PLEG): container finished" podID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerID="a2721eb0c85a1081c47fb5c445a2ecfc96774ff7a97d0d0d32c4962dc0c42219" exitCode=0 Oct 06 08:54:32 crc kubenswrapper[4989]: I1006 08:54:32.957895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerDied","Data":"a2721eb0c85a1081c47fb5c445a2ecfc96774ff7a97d0d0d32c4962dc0c42219"} Oct 06 08:54:33 crc kubenswrapper[4989]: I1006 08:54:33.966419 4989 generic.go:334] "Generic (PLEG): container finished" podID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerID="3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9" exitCode=0 Oct 06 08:54:33 crc kubenswrapper[4989]: I1006 08:54:33.967273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerDied","Data":"3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9"} Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.213605 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.250950 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util\") pod \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.251017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle\") pod \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.251052 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4rhm\" (UniqueName: \"kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm\") pod \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\" (UID: \"93fa8693-a1bc-4438-8a3c-318c6374f1b4\") " Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.252310 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle" (OuterVolumeSpecName: "bundle") pod "93fa8693-a1bc-4438-8a3c-318c6374f1b4" (UID: "93fa8693-a1bc-4438-8a3c-318c6374f1b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.257012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm" (OuterVolumeSpecName: "kube-api-access-n4rhm") pod "93fa8693-a1bc-4438-8a3c-318c6374f1b4" (UID: "93fa8693-a1bc-4438-8a3c-318c6374f1b4"). InnerVolumeSpecName "kube-api-access-n4rhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.265890 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util" (OuterVolumeSpecName: "util") pod "93fa8693-a1bc-4438-8a3c-318c6374f1b4" (UID: "93fa8693-a1bc-4438-8a3c-318c6374f1b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.353382 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.353484 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4rhm\" (UniqueName: \"kubernetes.io/projected/93fa8693-a1bc-4438-8a3c-318c6374f1b4-kube-api-access-n4rhm\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.353587 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/93fa8693-a1bc-4438-8a3c-318c6374f1b4-util\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.972903 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" event={"ID":"93fa8693-a1bc-4438-8a3c-318c6374f1b4","Type":"ContainerDied","Data":"29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0"} Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.973191 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b35310946cddbf083155970e04bfa06368ad7596260b4f820e4a5db096d3b0" Oct 06 08:54:34 crc kubenswrapper[4989]: I1006 08:54:34.972991 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g" Oct 06 08:54:35 crc kubenswrapper[4989]: I1006 08:54:35.980368 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerStarted","Data":"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6"} Oct 06 08:54:36 crc kubenswrapper[4989]: I1006 08:54:36.002519 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-klw6n" podStartSLOduration=3.034179312 podStartE2EDuration="6.002503251s" podCreationTimestamp="2025-10-06 08:54:30 +0000 UTC" firstStartedPulling="2025-10-06 08:54:31.950495685 +0000 UTC m=+922.740521265" lastFinishedPulling="2025-10-06 08:54:34.918819624 +0000 UTC m=+925.708845204" observedRunningTime="2025-10-06 08:54:36.001082979 +0000 UTC m=+926.791108589" watchObservedRunningTime="2025-10-06 08:54:36.002503251 +0000 UTC m=+926.792528831" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.065811 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-2btck"] Oct 06 08:54:38 crc kubenswrapper[4989]: E1006 08:54:38.066287 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="pull" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.066298 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="pull" Oct 06 08:54:38 crc kubenswrapper[4989]: E1006 08:54:38.066319 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="util" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.066325 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="util" Oct 06 08:54:38 crc kubenswrapper[4989]: E1006 08:54:38.066332 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="extract" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.066338 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="extract" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.066417 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="93fa8693-a1bc-4438-8a3c-318c6374f1b4" containerName="extract" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.066790 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.069162 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.069217 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.069347 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jmbbg" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.077590 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-2btck"] Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.100527 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmwzf\" (UniqueName: \"kubernetes.io/projected/4f4c242d-e5b1-4091-9960-9c7e7141f04f-kube-api-access-kmwzf\") pod \"nmstate-operator-858ddd8f98-2btck\" (UID: \"4f4c242d-e5b1-4091-9960-9c7e7141f04f\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.201643 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmwzf\" (UniqueName: \"kubernetes.io/projected/4f4c242d-e5b1-4091-9960-9c7e7141f04f-kube-api-access-kmwzf\") pod \"nmstate-operator-858ddd8f98-2btck\" (UID: \"4f4c242d-e5b1-4091-9960-9c7e7141f04f\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.224307 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmwzf\" (UniqueName: \"kubernetes.io/projected/4f4c242d-e5b1-4091-9960-9c7e7141f04f-kube-api-access-kmwzf\") pod \"nmstate-operator-858ddd8f98-2btck\" (UID: \"4f4c242d-e5b1-4091-9960-9c7e7141f04f\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.379784 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.774259 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-2btck"] Oct 06 08:54:38 crc kubenswrapper[4989]: W1006 08:54:38.782046 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f4c242d_e5b1_4091_9960_9c7e7141f04f.slice/crio-4353dbed5bd1ebc9d19d05dde3389ce8172f55a5a16fbae7fe57dc3fd5b34ce6 WatchSource:0}: Error finding container 4353dbed5bd1ebc9d19d05dde3389ce8172f55a5a16fbae7fe57dc3fd5b34ce6: Status 404 returned error can't find the container with id 4353dbed5bd1ebc9d19d05dde3389ce8172f55a5a16fbae7fe57dc3fd5b34ce6 Oct 06 08:54:38 crc kubenswrapper[4989]: I1006 08:54:38.998204 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" event={"ID":"4f4c242d-e5b1-4091-9960-9c7e7141f04f","Type":"ContainerStarted","Data":"4353dbed5bd1ebc9d19d05dde3389ce8172f55a5a16fbae7fe57dc3fd5b34ce6"} Oct 06 08:54:40 crc kubenswrapper[4989]: I1006 08:54:40.461563 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:40 crc kubenswrapper[4989]: I1006 08:54:40.461936 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:40 crc kubenswrapper[4989]: I1006 08:54:40.505361 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:41 crc kubenswrapper[4989]: I1006 08:54:41.070064 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:42 crc kubenswrapper[4989]: I1006 08:54:42.019249 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" event={"ID":"4f4c242d-e5b1-4091-9960-9c7e7141f04f","Type":"ContainerStarted","Data":"d8cdb2aeb6ac90422bb2f6c5633129ccabc5ca16a4676472a2e5581956471c97"} Oct 06 08:54:42 crc kubenswrapper[4989]: I1006 08:54:42.052203 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-2btck" podStartSLOduration=1.9165535120000001 podStartE2EDuration="4.052174027s" podCreationTimestamp="2025-10-06 08:54:38 +0000 UTC" firstStartedPulling="2025-10-06 08:54:38.784759741 +0000 UTC m=+929.574785331" lastFinishedPulling="2025-10-06 08:54:40.920380266 +0000 UTC m=+931.710405846" observedRunningTime="2025-10-06 08:54:42.04651538 +0000 UTC m=+932.836540970" watchObservedRunningTime="2025-10-06 08:54:42.052174027 +0000 UTC m=+932.842199617" Oct 06 08:54:42 crc kubenswrapper[4989]: I1006 08:54:42.932154 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.025804 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-klw6n" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="registry-server" containerID="cri-o://d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6" gracePeriod=2 Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.388530 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.461145 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhm66\" (UniqueName: \"kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66\") pod \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.461221 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content\") pod \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.461279 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities\") pod \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\" (UID: \"8d2ddb93-a5a3-4eca-8b55-23b60be2f788\") " Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.463571 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities" (OuterVolumeSpecName: "utilities") pod "8d2ddb93-a5a3-4eca-8b55-23b60be2f788" (UID: "8d2ddb93-a5a3-4eca-8b55-23b60be2f788"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.469107 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66" (OuterVolumeSpecName: "kube-api-access-xhm66") pod "8d2ddb93-a5a3-4eca-8b55-23b60be2f788" (UID: "8d2ddb93-a5a3-4eca-8b55-23b60be2f788"). InnerVolumeSpecName "kube-api-access-xhm66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.546572 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d2ddb93-a5a3-4eca-8b55-23b60be2f788" (UID: "8d2ddb93-a5a3-4eca-8b55-23b60be2f788"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.563684 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhm66\" (UniqueName: \"kubernetes.io/projected/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-kube-api-access-xhm66\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.563720 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:43 crc kubenswrapper[4989]: I1006 08:54:43.563732 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ddb93-a5a3-4eca-8b55-23b60be2f788-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.033449 4989 generic.go:334] "Generic (PLEG): container finished" podID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerID="d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6" exitCode=0 Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.033513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerDied","Data":"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6"} Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.033549 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw6n" event={"ID":"8d2ddb93-a5a3-4eca-8b55-23b60be2f788","Type":"ContainerDied","Data":"250614d6e34bbd29277dc604aad606c77d6e9d7b499cebc71afafb1c1da82105"} Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.033544 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw6n" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.033566 4989 scope.go:117] "RemoveContainer" containerID="d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.053175 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.068340 4989 scope.go:117] "RemoveContainer" containerID="3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.069703 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-klw6n"] Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.091390 4989 scope.go:117] "RemoveContainer" containerID="f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.117425 4989 scope.go:117] "RemoveContainer" containerID="d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6" Oct 06 08:54:44 crc kubenswrapper[4989]: E1006 08:54:44.117921 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6\": container with ID starting with d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6 not found: ID does not exist" containerID="d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.118046 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6"} err="failed to get container status \"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6\": rpc error: code = NotFound desc = could not find container \"d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6\": container with ID starting with d894fc5317bcd9ebc06a6c6d37944941bb8ef9d746ae7234df249127b94659a6 not found: ID does not exist" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.118146 4989 scope.go:117] "RemoveContainer" containerID="3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9" Oct 06 08:54:44 crc kubenswrapper[4989]: E1006 08:54:44.118700 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9\": container with ID starting with 3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9 not found: ID does not exist" containerID="3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.118825 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9"} err="failed to get container status \"3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9\": rpc error: code = NotFound desc = could not find container \"3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9\": container with ID starting with 3e3bfd190094abffc46ca0a41c3c9080a5ab14692bb03ba4c4558440db85b6e9 not found: ID does not exist" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.118933 4989 scope.go:117] "RemoveContainer" containerID="f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd" Oct 06 08:54:44 crc kubenswrapper[4989]: E1006 08:54:44.119296 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd\": container with ID starting with f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd not found: ID does not exist" containerID="f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd" Oct 06 08:54:44 crc kubenswrapper[4989]: I1006 08:54:44.119329 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd"} err="failed to get container status \"f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd\": rpc error: code = NotFound desc = could not find container \"f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd\": container with ID starting with f89d4ce1225b7c6a9dd0996bcfd0de6e56ef883318fd80966551c0d770a334fd not found: ID does not exist" Oct 06 08:54:45 crc kubenswrapper[4989]: I1006 08:54:45.948877 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" path="/var/lib/kubelet/pods/8d2ddb93-a5a3-4eca-8b55-23b60be2f788/volumes" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.017078 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx"] Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.017302 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="registry-server" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.017316 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="registry-server" Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.017329 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="extract-content" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.017338 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="extract-content" Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.017354 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="extract-utilities" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.017361 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="extract-utilities" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.017461 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2ddb93-a5a3-4eca-8b55-23b60be2f788" containerName="registry-server" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.018153 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.024493 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-b8p5h" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.032690 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.045563 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.046637 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.051404 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-j7w6r"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.052066 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.055209 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.066539 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118402 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-dbus-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118696 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rjq5\" (UniqueName: \"kubernetes.io/projected/734fb7db-406a-47af-ae7b-5e346452ce57-kube-api-access-6rjq5\") pod \"nmstate-metrics-fdff9cb8d-85qpx\" (UID: \"734fb7db-406a-47af-ae7b-5e346452ce57\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118720 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118743 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-ovs-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118773 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfm9m\" (UniqueName: \"kubernetes.io/projected/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-kube-api-access-wfm9m\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118790 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-nmstate-lock\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.118810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28d49\" (UniqueName: \"kubernetes.io/projected/fc10f366-5738-466e-96ac-352e579427aa-kube-api-access-28d49\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.190155 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.190929 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.203611 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.203645 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-pnddj" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.203814 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.207970 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219526 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-ovs-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219594 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slzkw\" (UniqueName: \"kubernetes.io/projected/2ca0d980-4342-41a1-a47a-d8c55d87254c-kube-api-access-slzkw\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219627 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfm9m\" (UniqueName: \"kubernetes.io/projected/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-kube-api-access-wfm9m\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219666 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-nmstate-lock\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28d49\" (UniqueName: \"kubernetes.io/projected/fc10f366-5738-466e-96ac-352e579427aa-kube-api-access-28d49\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219763 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-dbus-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219808 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rjq5\" (UniqueName: \"kubernetes.io/projected/734fb7db-406a-47af-ae7b-5e346452ce57-kube-api-access-6rjq5\") pod \"nmstate-metrics-fdff9cb8d-85qpx\" (UID: \"734fb7db-406a-47af-ae7b-5e346452ce57\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219830 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219851 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2ca0d980-4342-41a1-a47a-d8c55d87254c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.219936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-ovs-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.220253 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-nmstate-lock\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.220741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-dbus-socket\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.220971 4989 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.221020 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair podName:fc10f366-5738-466e-96ac-352e579427aa nodeName:}" failed. No retries permitted until 2025-10-06 08:54:48.721000273 +0000 UTC m=+939.511025853 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair") pod "nmstate-webhook-6cdbc54649-sx7g9" (UID: "fc10f366-5738-466e-96ac-352e579427aa") : secret "openshift-nmstate-webhook" not found Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.243138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rjq5\" (UniqueName: \"kubernetes.io/projected/734fb7db-406a-47af-ae7b-5e346452ce57-kube-api-access-6rjq5\") pod \"nmstate-metrics-fdff9cb8d-85qpx\" (UID: \"734fb7db-406a-47af-ae7b-5e346452ce57\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.244013 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28d49\" (UniqueName: \"kubernetes.io/projected/fc10f366-5738-466e-96ac-352e579427aa-kube-api-access-28d49\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.244148 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfm9m\" (UniqueName: \"kubernetes.io/projected/e62c6594-51c2-42d8-9b9e-d8ed53a60be4-kube-api-access-wfm9m\") pod \"nmstate-handler-j7w6r\" (UID: \"e62c6594-51c2-42d8-9b9e-d8ed53a60be4\") " pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.321150 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slzkw\" (UniqueName: \"kubernetes.io/projected/2ca0d980-4342-41a1-a47a-d8c55d87254c-kube-api-access-slzkw\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.321403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.321585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2ca0d980-4342-41a1-a47a-d8c55d87254c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.321615 4989 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 06 08:54:48 crc kubenswrapper[4989]: E1006 08:54:48.321853 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert podName:2ca0d980-4342-41a1-a47a-d8c55d87254c nodeName:}" failed. No retries permitted until 2025-10-06 08:54:48.821835538 +0000 UTC m=+939.611861118 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-fglz7" (UID: "2ca0d980-4342-41a1-a47a-d8c55d87254c") : secret "plugin-serving-cert" not found Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.322732 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2ca0d980-4342-41a1-a47a-d8c55d87254c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.334369 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.351820 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slzkw\" (UniqueName: \"kubernetes.io/projected/2ca0d980-4342-41a1-a47a-d8c55d87254c-kube-api-access-slzkw\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.377941 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.413237 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-67ff44877f-gkp52"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.414169 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-trusted-ca-bundle\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426313 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426331 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-service-ca\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426361 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-oauth-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426383 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426422 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbclp\" (UniqueName: \"kubernetes.io/projected/052d1af0-8fa0-444b-b2e9-82ae3463c09f-kube-api-access-cbclp\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.426444 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-oauth-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.463191 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67ff44877f-gkp52"] Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529251 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-oauth-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529593 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529647 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbclp\" (UniqueName: \"kubernetes.io/projected/052d1af0-8fa0-444b-b2e9-82ae3463c09f-kube-api-access-cbclp\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529685 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-oauth-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529719 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-trusted-ca-bundle\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529744 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.529758 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-service-ca\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.531785 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.533470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-trusted-ca-bundle\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.535007 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-oauth-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.535051 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-serving-cert\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.535065 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/052d1af0-8fa0-444b-b2e9-82ae3463c09f-console-oauth-config\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.536126 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/052d1af0-8fa0-444b-b2e9-82ae3463c09f-service-ca\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.551800 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbclp\" (UniqueName: \"kubernetes.io/projected/052d1af0-8fa0-444b-b2e9-82ae3463c09f-kube-api-access-cbclp\") pod \"console-67ff44877f-gkp52\" (UID: \"052d1af0-8fa0-444b-b2e9-82ae3463c09f\") " pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.626811 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx"] Oct 06 08:54:48 crc kubenswrapper[4989]: W1006 08:54:48.629124 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod734fb7db_406a_47af_ae7b_5e346452ce57.slice/crio-b57bfb655362b35235e85b734560f164c2d5d48b919c037089c75647fda3fc18 WatchSource:0}: Error finding container b57bfb655362b35235e85b734560f164c2d5d48b919c037089c75647fda3fc18: Status 404 returned error can't find the container with id b57bfb655362b35235e85b734560f164c2d5d48b919c037089c75647fda3fc18 Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.731786 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.736295 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fc10f366-5738-466e-96ac-352e579427aa-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-sx7g9\" (UID: \"fc10f366-5738-466e-96ac-352e579427aa\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.789057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.832623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.836297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2ca0d980-4342-41a1-a47a-d8c55d87254c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fglz7\" (UID: \"2ca0d980-4342-41a1-a47a-d8c55d87254c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:48 crc kubenswrapper[4989]: I1006 08:54:48.968172 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.001334 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67ff44877f-gkp52"] Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.067375 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" event={"ID":"734fb7db-406a-47af-ae7b-5e346452ce57","Type":"ContainerStarted","Data":"b57bfb655362b35235e85b734560f164c2d5d48b919c037089c75647fda3fc18"} Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.068340 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j7w6r" event={"ID":"e62c6594-51c2-42d8-9b9e-d8ed53a60be4","Type":"ContainerStarted","Data":"66ccadae6d5688549a2de75f45721e7e24e4fd177773b5c5bce5d803b9947219"} Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.069833 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67ff44877f-gkp52" event={"ID":"052d1af0-8fa0-444b-b2e9-82ae3463c09f","Type":"ContainerStarted","Data":"6aed3f4a1ee436f471c92eb592bdeec9654c64482b0c9257656ae4a229a89b00"} Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.107693 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.150377 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9"] Oct 06 08:54:49 crc kubenswrapper[4989]: W1006 08:54:49.157501 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc10f366_5738_466e_96ac_352e579427aa.slice/crio-7f5f200244d4a209e7b66a9c2a94cba829addd40d8044c80cb90d7b80b7ff2de WatchSource:0}: Error finding container 7f5f200244d4a209e7b66a9c2a94cba829addd40d8044c80cb90d7b80b7ff2de: Status 404 returned error can't find the container with id 7f5f200244d4a209e7b66a9c2a94cba829addd40d8044c80cb90d7b80b7ff2de Oct 06 08:54:49 crc kubenswrapper[4989]: I1006 08:54:49.321182 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7"] Oct 06 08:54:50 crc kubenswrapper[4989]: I1006 08:54:50.077537 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67ff44877f-gkp52" event={"ID":"052d1af0-8fa0-444b-b2e9-82ae3463c09f","Type":"ContainerStarted","Data":"7d805ded93f403dec9be6a0e1aa5c22f5178b2c7a07bcb815314fd5cbc79c91a"} Oct 06 08:54:50 crc kubenswrapper[4989]: I1006 08:54:50.080690 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" event={"ID":"fc10f366-5738-466e-96ac-352e579427aa","Type":"ContainerStarted","Data":"7f5f200244d4a209e7b66a9c2a94cba829addd40d8044c80cb90d7b80b7ff2de"} Oct 06 08:54:50 crc kubenswrapper[4989]: I1006 08:54:50.082186 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" event={"ID":"2ca0d980-4342-41a1-a47a-d8c55d87254c","Type":"ContainerStarted","Data":"480774781af700e19689810b2d778ada0929ac776eafdce1bec91ece433dc02f"} Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.096321 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" event={"ID":"734fb7db-406a-47af-ae7b-5e346452ce57","Type":"ContainerStarted","Data":"57f578597b94335c1702065688cf6292746559fff888433631f003e515165019"} Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.100097 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-j7w6r" event={"ID":"e62c6594-51c2-42d8-9b9e-d8ed53a60be4","Type":"ContainerStarted","Data":"8c4186a41f8faff5ee5c7a4df14d24fec4594cd98219347fe71ec13025113740"} Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.100190 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.103718 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" event={"ID":"fc10f366-5738-466e-96ac-352e579427aa","Type":"ContainerStarted","Data":"1f366c0c3366c1a450d7cd1c72bcb7b6aa30a8e4de6a2bb21d4ff80b80b42ee2"} Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.103855 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.117486 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-j7w6r" podStartSLOduration=1.3227618749999999 podStartE2EDuration="4.117450953s" podCreationTimestamp="2025-10-06 08:54:48 +0000 UTC" firstStartedPulling="2025-10-06 08:54:48.472923422 +0000 UTC m=+939.262949002" lastFinishedPulling="2025-10-06 08:54:51.26761249 +0000 UTC m=+942.057638080" observedRunningTime="2025-10-06 08:54:52.113758624 +0000 UTC m=+942.903784214" watchObservedRunningTime="2025-10-06 08:54:52.117450953 +0000 UTC m=+942.907476533" Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.117787 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-67ff44877f-gkp52" podStartSLOduration=4.117782893 podStartE2EDuration="4.117782893s" podCreationTimestamp="2025-10-06 08:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:54:50.09703887 +0000 UTC m=+940.887064470" watchObservedRunningTime="2025-10-06 08:54:52.117782893 +0000 UTC m=+942.907808473" Oct 06 08:54:52 crc kubenswrapper[4989]: I1006 08:54:52.133306 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" podStartSLOduration=2.026539092 podStartE2EDuration="4.133289542s" podCreationTimestamp="2025-10-06 08:54:48 +0000 UTC" firstStartedPulling="2025-10-06 08:54:49.159865681 +0000 UTC m=+939.949891261" lastFinishedPulling="2025-10-06 08:54:51.266616091 +0000 UTC m=+942.056641711" observedRunningTime="2025-10-06 08:54:52.125918964 +0000 UTC m=+942.915944554" watchObservedRunningTime="2025-10-06 08:54:52.133289542 +0000 UTC m=+942.923315122" Oct 06 08:54:53 crc kubenswrapper[4989]: I1006 08:54:53.112568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" event={"ID":"2ca0d980-4342-41a1-a47a-d8c55d87254c","Type":"ContainerStarted","Data":"acbb12ce4af885ea9e9b910034e48f2a27d642798ffd2759bb0fdf62748b204a"} Oct 06 08:54:53 crc kubenswrapper[4989]: I1006 08:54:53.127564 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fglz7" podStartSLOduration=2.194255997 podStartE2EDuration="5.12753583s" podCreationTimestamp="2025-10-06 08:54:48 +0000 UTC" firstStartedPulling="2025-10-06 08:54:49.330426981 +0000 UTC m=+940.120452571" lastFinishedPulling="2025-10-06 08:54:52.263706824 +0000 UTC m=+943.053732404" observedRunningTime="2025-10-06 08:54:53.125856181 +0000 UTC m=+943.915881771" watchObservedRunningTime="2025-10-06 08:54:53.12753583 +0000 UTC m=+943.917561450" Oct 06 08:54:55 crc kubenswrapper[4989]: I1006 08:54:55.126197 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" event={"ID":"734fb7db-406a-47af-ae7b-5e346452ce57","Type":"ContainerStarted","Data":"2a0ab40ffbbcee94188cc22b5c8d23f3d604b6127d6dbe68f11535707e986dbb"} Oct 06 08:54:55 crc kubenswrapper[4989]: I1006 08:54:55.151609 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-85qpx" podStartSLOduration=1.47895806 podStartE2EDuration="7.151578171s" podCreationTimestamp="2025-10-06 08:54:48 +0000 UTC" firstStartedPulling="2025-10-06 08:54:48.631058174 +0000 UTC m=+939.421083754" lastFinishedPulling="2025-10-06 08:54:54.303678285 +0000 UTC m=+945.093703865" observedRunningTime="2025-10-06 08:54:55.150457818 +0000 UTC m=+945.940483438" watchObservedRunningTime="2025-10-06 08:54:55.151578171 +0000 UTC m=+945.941603801" Oct 06 08:54:58 crc kubenswrapper[4989]: I1006 08:54:58.400796 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-j7w6r" Oct 06 08:54:58 crc kubenswrapper[4989]: I1006 08:54:58.789978 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:58 crc kubenswrapper[4989]: I1006 08:54:58.790037 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:58 crc kubenswrapper[4989]: I1006 08:54:58.794615 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:59 crc kubenswrapper[4989]: I1006 08:54:59.154044 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-67ff44877f-gkp52" Oct 06 08:54:59 crc kubenswrapper[4989]: I1006 08:54:59.217081 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:55:08 crc kubenswrapper[4989]: I1006 08:55:08.977028 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-sx7g9" Oct 06 08:55:11 crc kubenswrapper[4989]: I1006 08:55:11.891085 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:11 crc kubenswrapper[4989]: I1006 08:55:11.892689 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:11 crc kubenswrapper[4989]: I1006 08:55:11.904493 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.043510 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn82z\" (UniqueName: \"kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.043577 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.043635 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.145047 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn82z\" (UniqueName: \"kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.145108 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.145149 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.145671 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.146315 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.166450 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn82z\" (UniqueName: \"kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z\") pod \"community-operators-x4ddk\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.215112 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:12 crc kubenswrapper[4989]: I1006 08:55:12.477050 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:13 crc kubenswrapper[4989]: I1006 08:55:13.245592 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerID="58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733" exitCode=0 Oct 06 08:55:13 crc kubenswrapper[4989]: I1006 08:55:13.245638 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerDied","Data":"58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733"} Oct 06 08:55:13 crc kubenswrapper[4989]: I1006 08:55:13.245684 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerStarted","Data":"6dbef2bcdfc73e8a26ab80948be217fc0b02d750dcc470fb21a853a5224b5c1b"} Oct 06 08:55:15 crc kubenswrapper[4989]: I1006 08:55:15.260991 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerID="6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb" exitCode=0 Oct 06 08:55:15 crc kubenswrapper[4989]: I1006 08:55:15.261034 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerDied","Data":"6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb"} Oct 06 08:55:16 crc kubenswrapper[4989]: I1006 08:55:16.273747 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerStarted","Data":"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba"} Oct 06 08:55:16 crc kubenswrapper[4989]: I1006 08:55:16.300439 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x4ddk" podStartSLOduration=2.728858458 podStartE2EDuration="5.3004163s" podCreationTimestamp="2025-10-06 08:55:11 +0000 UTC" firstStartedPulling="2025-10-06 08:55:13.24799053 +0000 UTC m=+964.038016110" lastFinishedPulling="2025-10-06 08:55:15.819548372 +0000 UTC m=+966.609573952" observedRunningTime="2025-10-06 08:55:16.295177725 +0000 UTC m=+967.085203305" watchObservedRunningTime="2025-10-06 08:55:16.3004163 +0000 UTC m=+967.090441880" Oct 06 08:55:22 crc kubenswrapper[4989]: I1006 08:55:22.216827 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:22 crc kubenswrapper[4989]: I1006 08:55:22.217625 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:22 crc kubenswrapper[4989]: I1006 08:55:22.281493 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:22 crc kubenswrapper[4989]: I1006 08:55:22.389113 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:22 crc kubenswrapper[4989]: I1006 08:55:22.516343 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.176536 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9"] Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.177959 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.180015 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.189401 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.189819 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srwnj\" (UniqueName: \"kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.189877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.190415 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9"] Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.290627 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srwnj\" (UniqueName: \"kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.290721 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.290780 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.291588 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.291619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.311451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srwnj\" (UniqueName: \"kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.499075 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:23 crc kubenswrapper[4989]: I1006 08:55:23.727394 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9"] Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.258163 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wwr9w" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerName="console" containerID="cri-o://912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4" gracePeriod=15 Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.319018 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerID="5218b6d7edc405bd56e847e8579c327b0eeda1ed5917eb1d766db8832a81e220" exitCode=0 Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.319109 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" event={"ID":"4f4d3ca6-b70d-465c-9a76-a98e046542c4","Type":"ContainerDied","Data":"5218b6d7edc405bd56e847e8579c327b0eeda1ed5917eb1d766db8832a81e220"} Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.319457 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" event={"ID":"4f4d3ca6-b70d-465c-9a76-a98e046542c4","Type":"ContainerStarted","Data":"a0a9607357fbc854a6faed8e60201bebd353b265e61a2645eab862c63178d04b"} Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.319641 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x4ddk" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="registry-server" containerID="cri-o://f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba" gracePeriod=2 Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.656763 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wwr9w_b3a7037e-5b9e-4440-a24c-aaca9364fe0f/console/0.log" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.656829 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.665451 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.718876 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmgjk\" (UniqueName: \"kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.718974 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719021 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content\") pod \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719068 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719092 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719116 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719185 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn82z\" (UniqueName: \"kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z\") pod \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719255 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719295 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config\") pod \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\" (UID: \"b3a7037e-5b9e-4440-a24c-aaca9364fe0f\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.719314 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities\") pod \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\" (UID: \"1dd90244-6a0a-48fc-bb30-45cfff4c0ade\") " Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.721159 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config" (OuterVolumeSpecName: "console-config") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.721195 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca" (OuterVolumeSpecName: "service-ca") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.721288 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.721490 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities" (OuterVolumeSpecName: "utilities") pod "1dd90244-6a0a-48fc-bb30-45cfff4c0ade" (UID: "1dd90244-6a0a-48fc-bb30-45cfff4c0ade"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.722286 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.727862 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z" (OuterVolumeSpecName: "kube-api-access-mn82z") pod "1dd90244-6a0a-48fc-bb30-45cfff4c0ade" (UID: "1dd90244-6a0a-48fc-bb30-45cfff4c0ade"). InnerVolumeSpecName "kube-api-access-mn82z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.728204 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk" (OuterVolumeSpecName: "kube-api-access-tmgjk") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "kube-api-access-tmgjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.728253 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.730682 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b3a7037e-5b9e-4440-a24c-aaca9364fe0f" (UID: "b3a7037e-5b9e-4440-a24c-aaca9364fe0f"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822004 4989 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822034 4989 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822044 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822054 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmgjk\" (UniqueName: \"kubernetes.io/projected/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-kube-api-access-tmgjk\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822064 4989 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822074 4989 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-console-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822082 4989 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-service-ca\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822090 4989 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3a7037e-5b9e-4440-a24c-aaca9364fe0f-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:24 crc kubenswrapper[4989]: I1006 08:55:24.822098 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn82z\" (UniqueName: \"kubernetes.io/projected/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-kube-api-access-mn82z\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.327493 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wwr9w_b3a7037e-5b9e-4440-a24c-aaca9364fe0f/console/0.log" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.327966 4989 generic.go:334] "Generic (PLEG): container finished" podID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerID="912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4" exitCode=2 Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.328062 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wwr9w" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.328065 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wwr9w" event={"ID":"b3a7037e-5b9e-4440-a24c-aaca9364fe0f","Type":"ContainerDied","Data":"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4"} Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.328146 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wwr9w" event={"ID":"b3a7037e-5b9e-4440-a24c-aaca9364fe0f","Type":"ContainerDied","Data":"bccda34975faf5a32dbbc1c66d61eb9efe43b1172525f89ec9001fd8e05e9558"} Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.328185 4989 scope.go:117] "RemoveContainer" containerID="912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.331618 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerID="f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba" exitCode=0 Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.331714 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerDied","Data":"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba"} Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.331765 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4ddk" event={"ID":"1dd90244-6a0a-48fc-bb30-45cfff4c0ade","Type":"ContainerDied","Data":"6dbef2bcdfc73e8a26ab80948be217fc0b02d750dcc470fb21a853a5224b5c1b"} Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.331846 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4ddk" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.352231 4989 scope.go:117] "RemoveContainer" containerID="912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4" Oct 06 08:55:25 crc kubenswrapper[4989]: E1006 08:55:25.353065 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4\": container with ID starting with 912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4 not found: ID does not exist" containerID="912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.353404 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4"} err="failed to get container status \"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4\": rpc error: code = NotFound desc = could not find container \"912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4\": container with ID starting with 912773976cc4aeb64fa9dd886d65fa9275d772fba540eadf7de4b840535b92d4 not found: ID does not exist" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.353447 4989 scope.go:117] "RemoveContainer" containerID="f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.377327 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.380348 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wwr9w"] Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.385009 4989 scope.go:117] "RemoveContainer" containerID="6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.413245 4989 scope.go:117] "RemoveContainer" containerID="58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.457245 4989 scope.go:117] "RemoveContainer" containerID="f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba" Oct 06 08:55:25 crc kubenswrapper[4989]: E1006 08:55:25.457731 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba\": container with ID starting with f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba not found: ID does not exist" containerID="f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.457771 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba"} err="failed to get container status \"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba\": rpc error: code = NotFound desc = could not find container \"f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba\": container with ID starting with f73c527b4265e440a83429da4ae24e99e51798f52abd4746e89be6345b4c82ba not found: ID does not exist" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.457797 4989 scope.go:117] "RemoveContainer" containerID="6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb" Oct 06 08:55:25 crc kubenswrapper[4989]: E1006 08:55:25.458310 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb\": container with ID starting with 6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb not found: ID does not exist" containerID="6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.458345 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb"} err="failed to get container status \"6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb\": rpc error: code = NotFound desc = could not find container \"6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb\": container with ID starting with 6906ac66de4af4be2fd052128f958d0f58d2236509c31bf27a67b290cb6ebecb not found: ID does not exist" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.458363 4989 scope.go:117] "RemoveContainer" containerID="58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733" Oct 06 08:55:25 crc kubenswrapper[4989]: E1006 08:55:25.458630 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733\": container with ID starting with 58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733 not found: ID does not exist" containerID="58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.458672 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733"} err="failed to get container status \"58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733\": rpc error: code = NotFound desc = could not find container \"58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733\": container with ID starting with 58ab13dad9ea9e01d1cb81ca18f8f5360838c318f8cb35cb9d5836a68ed17733 not found: ID does not exist" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.575735 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dd90244-6a0a-48fc-bb30-45cfff4c0ade" (UID: "1dd90244-6a0a-48fc-bb30-45cfff4c0ade"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.631045 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dd90244-6a0a-48fc-bb30-45cfff4c0ade-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.667046 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.672304 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x4ddk"] Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.942850 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" path="/var/lib/kubelet/pods/1dd90244-6a0a-48fc-bb30-45cfff4c0ade/volumes" Oct 06 08:55:25 crc kubenswrapper[4989]: I1006 08:55:25.943647 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" path="/var/lib/kubelet/pods/b3a7037e-5b9e-4440-a24c-aaca9364fe0f/volumes" Oct 06 08:55:27 crc kubenswrapper[4989]: I1006 08:55:27.349704 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerID="ea2ec3d15db988385d79567cfb20a20640b5766b29431e27a21164e83bfd4599" exitCode=0 Oct 06 08:55:27 crc kubenswrapper[4989]: I1006 08:55:27.349753 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" event={"ID":"4f4d3ca6-b70d-465c-9a76-a98e046542c4","Type":"ContainerDied","Data":"ea2ec3d15db988385d79567cfb20a20640b5766b29431e27a21164e83bfd4599"} Oct 06 08:55:28 crc kubenswrapper[4989]: I1006 08:55:28.358662 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerID="b07555f02ce1de9fd34be67198ab8de1d6e3045bc287a361379bb46cb0e7e281" exitCode=0 Oct 06 08:55:28 crc kubenswrapper[4989]: I1006 08:55:28.358806 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" event={"ID":"4f4d3ca6-b70d-465c-9a76-a98e046542c4","Type":"ContainerDied","Data":"b07555f02ce1de9fd34be67198ab8de1d6e3045bc287a361379bb46cb0e7e281"} Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.680202 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.787647 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srwnj\" (UniqueName: \"kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj\") pod \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.787714 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util\") pod \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.787776 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle\") pod \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\" (UID: \"4f4d3ca6-b70d-465c-9a76-a98e046542c4\") " Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.788700 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle" (OuterVolumeSpecName: "bundle") pod "4f4d3ca6-b70d-465c-9a76-a98e046542c4" (UID: "4f4d3ca6-b70d-465c-9a76-a98e046542c4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.794886 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj" (OuterVolumeSpecName: "kube-api-access-srwnj") pod "4f4d3ca6-b70d-465c-9a76-a98e046542c4" (UID: "4f4d3ca6-b70d-465c-9a76-a98e046542c4"). InnerVolumeSpecName "kube-api-access-srwnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.810989 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util" (OuterVolumeSpecName: "util") pod "4f4d3ca6-b70d-465c-9a76-a98e046542c4" (UID: "4f4d3ca6-b70d-465c-9a76-a98e046542c4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.888410 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.888451 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srwnj\" (UniqueName: \"kubernetes.io/projected/4f4d3ca6-b70d-465c-9a76-a98e046542c4-kube-api-access-srwnj\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:29 crc kubenswrapper[4989]: I1006 08:55:29.888466 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f4d3ca6-b70d-465c-9a76-a98e046542c4-util\") on node \"crc\" DevicePath \"\"" Oct 06 08:55:30 crc kubenswrapper[4989]: I1006 08:55:30.377164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" event={"ID":"4f4d3ca6-b70d-465c-9a76-a98e046542c4","Type":"ContainerDied","Data":"a0a9607357fbc854a6faed8e60201bebd353b265e61a2645eab862c63178d04b"} Oct 06 08:55:30 crc kubenswrapper[4989]: I1006 08:55:30.377251 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0a9607357fbc854a6faed8e60201bebd353b265e61a2645eab862c63178d04b" Oct 06 08:55:30 crc kubenswrapper[4989]: I1006 08:55:30.377398 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.084891 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn"] Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085719 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="extract" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085736 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="extract" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085750 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="extract-utilities" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085759 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="extract-utilities" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085768 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerName="console" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085775 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerName="console" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085790 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="registry-server" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085797 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="registry-server" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085808 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="pull" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085815 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="pull" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085828 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="extract-content" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085835 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="extract-content" Oct 06 08:55:39 crc kubenswrapper[4989]: E1006 08:55:39.085848 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="util" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085855 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="util" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085978 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f4d3ca6-b70d-465c-9a76-a98e046542c4" containerName="extract" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085990 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a7037e-5b9e-4440-a24c-aaca9364fe0f" containerName="console" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.085998 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dd90244-6a0a-48fc-bb30-45cfff4c0ade" containerName="registry-server" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.086444 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.088771 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.088818 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-jl4lw" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.089068 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.089174 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.089184 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.103705 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn"] Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.104472 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-webhook-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.104696 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-apiservice-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.104810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p66q7\" (UniqueName: \"kubernetes.io/projected/342ae3db-e559-4882-bc4b-6999d8cb1a13-kube-api-access-p66q7\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.206598 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-apiservice-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.206640 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p66q7\" (UniqueName: \"kubernetes.io/projected/342ae3db-e559-4882-bc4b-6999d8cb1a13-kube-api-access-p66q7\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.206690 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-webhook-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.212244 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-apiservice-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.212583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/342ae3db-e559-4882-bc4b-6999d8cb1a13-webhook-cert\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.226236 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p66q7\" (UniqueName: \"kubernetes.io/projected/342ae3db-e559-4882-bc4b-6999d8cb1a13-kube-api-access-p66q7\") pod \"metallb-operator-controller-manager-745c6df699-vqlnn\" (UID: \"342ae3db-e559-4882-bc4b-6999d8cb1a13\") " pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.401683 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf"] Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.402388 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.404463 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.406007 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.406037 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pv5bt" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.409328 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-apiservice-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.409550 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-webhook-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.409713 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mpf2\" (UniqueName: \"kubernetes.io/projected/9ff68b6e-9a02-4f31-8111-564bae539f3a-kube-api-access-8mpf2\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.421529 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf"] Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.443916 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.510970 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-apiservice-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.511513 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-webhook-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.511714 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mpf2\" (UniqueName: \"kubernetes.io/projected/9ff68b6e-9a02-4f31-8111-564bae539f3a-kube-api-access-8mpf2\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.523374 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-webhook-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.523399 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ff68b6e-9a02-4f31-8111-564bae539f3a-apiservice-cert\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.537556 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mpf2\" (UniqueName: \"kubernetes.io/projected/9ff68b6e-9a02-4f31-8111-564bae539f3a-kube-api-access-8mpf2\") pod \"metallb-operator-webhook-server-59f68d59f5-cv5nf\" (UID: \"9ff68b6e-9a02-4f31-8111-564bae539f3a\") " pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.664293 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn"] Oct 06 08:55:39 crc kubenswrapper[4989]: W1006 08:55:39.672038 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod342ae3db_e559_4882_bc4b_6999d8cb1a13.slice/crio-51e37d64f4cff93fe2683d8ea05f78d2f4d78252aeaa1fd45cafd90c40f637a9 WatchSource:0}: Error finding container 51e37d64f4cff93fe2683d8ea05f78d2f4d78252aeaa1fd45cafd90c40f637a9: Status 404 returned error can't find the container with id 51e37d64f4cff93fe2683d8ea05f78d2f4d78252aeaa1fd45cafd90c40f637a9 Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.725779 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:39 crc kubenswrapper[4989]: I1006 08:55:39.925450 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf"] Oct 06 08:55:39 crc kubenswrapper[4989]: W1006 08:55:39.935151 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ff68b6e_9a02_4f31_8111_564bae539f3a.slice/crio-68cd3aa5a11d15aeff0fe7a543fb66982abdbc7a5067eed5d5b9031b1722426f WatchSource:0}: Error finding container 68cd3aa5a11d15aeff0fe7a543fb66982abdbc7a5067eed5d5b9031b1722426f: Status 404 returned error can't find the container with id 68cd3aa5a11d15aeff0fe7a543fb66982abdbc7a5067eed5d5b9031b1722426f Oct 06 08:55:40 crc kubenswrapper[4989]: I1006 08:55:40.436925 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" event={"ID":"342ae3db-e559-4882-bc4b-6999d8cb1a13","Type":"ContainerStarted","Data":"51e37d64f4cff93fe2683d8ea05f78d2f4d78252aeaa1fd45cafd90c40f637a9"} Oct 06 08:55:40 crc kubenswrapper[4989]: I1006 08:55:40.437972 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" event={"ID":"9ff68b6e-9a02-4f31-8111-564bae539f3a","Type":"ContainerStarted","Data":"68cd3aa5a11d15aeff0fe7a543fb66982abdbc7a5067eed5d5b9031b1722426f"} Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.466297 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" event={"ID":"9ff68b6e-9a02-4f31-8111-564bae539f3a","Type":"ContainerStarted","Data":"5441a41b05dbe27e2bd45db1e8cbf216fedcf545c6d3c6e1ddde8ed70b2631b9"} Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.466922 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.470435 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" event={"ID":"342ae3db-e559-4882-bc4b-6999d8cb1a13","Type":"ContainerStarted","Data":"04e6bfa464110506d2597f0b3ebbe53d057f10572ee9441a8aed875f5bad40cd"} Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.470761 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.492266 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" podStartSLOduration=1.37584898 podStartE2EDuration="6.492245974s" podCreationTimestamp="2025-10-06 08:55:39 +0000 UTC" firstStartedPulling="2025-10-06 08:55:39.939451888 +0000 UTC m=+990.729477488" lastFinishedPulling="2025-10-06 08:55:45.055848902 +0000 UTC m=+995.845874482" observedRunningTime="2025-10-06 08:55:45.489698208 +0000 UTC m=+996.279723788" watchObservedRunningTime="2025-10-06 08:55:45.492245974 +0000 UTC m=+996.282271554" Oct 06 08:55:45 crc kubenswrapper[4989]: I1006 08:55:45.515750 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" podStartSLOduration=1.149607343 podStartE2EDuration="6.515733657s" podCreationTimestamp="2025-10-06 08:55:39 +0000 UTC" firstStartedPulling="2025-10-06 08:55:39.675565674 +0000 UTC m=+990.465591254" lastFinishedPulling="2025-10-06 08:55:45.041691988 +0000 UTC m=+995.831717568" observedRunningTime="2025-10-06 08:55:45.512033727 +0000 UTC m=+996.302059307" watchObservedRunningTime="2025-10-06 08:55:45.515733657 +0000 UTC m=+996.305759237" Oct 06 08:55:59 crc kubenswrapper[4989]: I1006 08:55:59.738367 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-59f68d59f5-cv5nf" Oct 06 08:56:03 crc kubenswrapper[4989]: I1006 08:56:03.935472 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:56:03 crc kubenswrapper[4989]: I1006 08:56:03.935867 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:56:19 crc kubenswrapper[4989]: I1006 08:56:19.447830 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-745c6df699-vqlnn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.165347 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-gxdgx"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.168123 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.173836 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.173940 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-22pjg" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.173995 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.178375 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-bs986"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.179168 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.181826 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.193092 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-bs986"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.259354 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-sqqsn"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.261331 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.264566 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.264626 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.264566 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.264834 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-928mm" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.269175 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-cdb7w"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.270462 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.274242 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.293720 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cdb7w"] Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a02ef7c3-577f-4284-88c6-ca16b3b335b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337582 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-reloader\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337613 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-conf\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337632 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mzd9\" (UniqueName: \"kubernetes.io/projected/a02ef7c3-577f-4284-88c6-ca16b3b335b8-kube-api-access-9mzd9\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337683 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38c9054e-238b-4eef-8f71-ec00117efac9-frr-startup\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337717 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmt6b\" (UniqueName: \"kubernetes.io/projected/38c9054e-238b-4eef-8f71-ec00117efac9-kube-api-access-bmt6b\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337914 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-metrics\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337951 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-sockets\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.337974 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c9054e-238b-4eef-8f71-ec00117efac9-metrics-certs\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439378 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-conf\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mzd9\" (UniqueName: \"kubernetes.io/projected/a02ef7c3-577f-4284-88c6-ca16b3b335b8-kube-api-access-9mzd9\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439453 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwjf8\" (UniqueName: \"kubernetes.io/projected/c021ef53-6c43-47fe-8429-f374626d4611-kube-api-access-hwjf8\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38c9054e-238b-4eef-8f71-ec00117efac9-frr-startup\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439502 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-cert\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439530 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmt6b\" (UniqueName: \"kubernetes.io/projected/38c9054e-238b-4eef-8f71-ec00117efac9-kube-api-access-bmt6b\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439552 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.439581 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-metrics\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-conf\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440396 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-sockets\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c9054e-238b-4eef-8f71-ec00117efac9-metrics-certs\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440476 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95h86\" (UniqueName: \"kubernetes.io/projected/cfd98297-3ba6-455c-b039-8fb182d89aa0-kube-api-access-95h86\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440500 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-frr-sockets\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440426 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-metrics\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440559 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a02ef7c3-577f-4284-88c6-ca16b3b335b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440638 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-reloader\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c021ef53-6c43-47fe-8429-f374626d4611-metallb-excludel2\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440725 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-metrics-certs\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440784 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-metrics-certs\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.440916 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38c9054e-238b-4eef-8f71-ec00117efac9-reloader\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.441011 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38c9054e-238b-4eef-8f71-ec00117efac9-frr-startup\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.446491 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38c9054e-238b-4eef-8f71-ec00117efac9-metrics-certs\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.453911 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a02ef7c3-577f-4284-88c6-ca16b3b335b8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.456256 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mzd9\" (UniqueName: \"kubernetes.io/projected/a02ef7c3-577f-4284-88c6-ca16b3b335b8-kube-api-access-9mzd9\") pod \"frr-k8s-webhook-server-64bf5d555-bs986\" (UID: \"a02ef7c3-577f-4284-88c6-ca16b3b335b8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.456409 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmt6b\" (UniqueName: \"kubernetes.io/projected/38c9054e-238b-4eef-8f71-ec00117efac9-kube-api-access-bmt6b\") pod \"frr-k8s-gxdgx\" (UID: \"38c9054e-238b-4eef-8f71-ec00117efac9\") " pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.485412 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.492819 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542322 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c021ef53-6c43-47fe-8429-f374626d4611-metallb-excludel2\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542366 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-metrics-certs\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542392 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-metrics-certs\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542427 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwjf8\" (UniqueName: \"kubernetes.io/projected/c021ef53-6c43-47fe-8429-f374626d4611-kube-api-access-hwjf8\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542461 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-cert\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542508 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.542554 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95h86\" (UniqueName: \"kubernetes.io/projected/cfd98297-3ba6-455c-b039-8fb182d89aa0-kube-api-access-95h86\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: E1006 08:56:20.543044 4989 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 06 08:56:20 crc kubenswrapper[4989]: E1006 08:56:20.543118 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist podName:c021ef53-6c43-47fe-8429-f374626d4611 nodeName:}" failed. No retries permitted until 2025-10-06 08:56:21.0430934 +0000 UTC m=+1031.833118980 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist") pod "speaker-sqqsn" (UID: "c021ef53-6c43-47fe-8429-f374626d4611") : secret "metallb-memberlist" not found Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.544344 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c021ef53-6c43-47fe-8429-f374626d4611-metallb-excludel2\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.546131 4989 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.550640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-metrics-certs\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.556545 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfd98297-3ba6-455c-b039-8fb182d89aa0-cert\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.559747 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-metrics-certs\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.563322 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwjf8\" (UniqueName: \"kubernetes.io/projected/c021ef53-6c43-47fe-8429-f374626d4611-kube-api-access-hwjf8\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.566895 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95h86\" (UniqueName: \"kubernetes.io/projected/cfd98297-3ba6-455c-b039-8fb182d89aa0-kube-api-access-95h86\") pod \"controller-68d546b9d8-cdb7w\" (UID: \"cfd98297-3ba6-455c-b039-8fb182d89aa0\") " pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.597300 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.678986 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"b0e311cdde499f1b4fa7d48342801acd13c98df92fcffe2bb3b1f762124cbe42"} Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.685956 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-bs986"] Oct 06 08:56:20 crc kubenswrapper[4989]: W1006 08:56:20.694736 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda02ef7c3_577f_4284_88c6_ca16b3b335b8.slice/crio-327e2b8114f847a4930eb214396476e9d408ce1e27465c5c7aed0e2cc0765d22 WatchSource:0}: Error finding container 327e2b8114f847a4930eb214396476e9d408ce1e27465c5c7aed0e2cc0765d22: Status 404 returned error can't find the container with id 327e2b8114f847a4930eb214396476e9d408ce1e27465c5c7aed0e2cc0765d22 Oct 06 08:56:20 crc kubenswrapper[4989]: I1006 08:56:20.771595 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cdb7w"] Oct 06 08:56:20 crc kubenswrapper[4989]: W1006 08:56:20.774327 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfd98297_3ba6_455c_b039_8fb182d89aa0.slice/crio-30f5f16b7fe47860312ab469de72a97f66331e431ce5a6a38dd2333a4f90f768 WatchSource:0}: Error finding container 30f5f16b7fe47860312ab469de72a97f66331e431ce5a6a38dd2333a4f90f768: Status 404 returned error can't find the container with id 30f5f16b7fe47860312ab469de72a97f66331e431ce5a6a38dd2333a4f90f768 Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.048723 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.054140 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c021ef53-6c43-47fe-8429-f374626d4611-memberlist\") pod \"speaker-sqqsn\" (UID: \"c021ef53-6c43-47fe-8429-f374626d4611\") " pod="metallb-system/speaker-sqqsn" Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.184130 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sqqsn" Oct 06 08:56:21 crc kubenswrapper[4989]: W1006 08:56:21.214778 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc021ef53_6c43_47fe_8429_f374626d4611.slice/crio-c51287d8076b5838fc79b9e7063d1e8418fdfa61a4db488b5fcdc3c9a5f7f319 WatchSource:0}: Error finding container c51287d8076b5838fc79b9e7063d1e8418fdfa61a4db488b5fcdc3c9a5f7f319: Status 404 returned error can't find the container with id c51287d8076b5838fc79b9e7063d1e8418fdfa61a4db488b5fcdc3c9a5f7f319 Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.685365 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" event={"ID":"a02ef7c3-577f-4284-88c6-ca16b3b335b8","Type":"ContainerStarted","Data":"327e2b8114f847a4930eb214396476e9d408ce1e27465c5c7aed0e2cc0765d22"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.687991 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sqqsn" event={"ID":"c021ef53-6c43-47fe-8429-f374626d4611","Type":"ContainerStarted","Data":"1fe18b795b398e607b2cd918c30e58079222be299e0457a6ee41dc83f18b915f"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.688019 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sqqsn" event={"ID":"c021ef53-6c43-47fe-8429-f374626d4611","Type":"ContainerStarted","Data":"c51287d8076b5838fc79b9e7063d1e8418fdfa61a4db488b5fcdc3c9a5f7f319"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.690073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdb7w" event={"ID":"cfd98297-3ba6-455c-b039-8fb182d89aa0","Type":"ContainerStarted","Data":"c149553b61409f5f892fd837f5453f457b5367bc9dff2567b1272b1f1b0c1c98"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.690124 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdb7w" event={"ID":"cfd98297-3ba6-455c-b039-8fb182d89aa0","Type":"ContainerStarted","Data":"70a6617d591a9d928a61a47b9eb76eff6b4a5370914cd52e00f5e0700b6448e0"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.690138 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cdb7w" event={"ID":"cfd98297-3ba6-455c-b039-8fb182d89aa0","Type":"ContainerStarted","Data":"30f5f16b7fe47860312ab469de72a97f66331e431ce5a6a38dd2333a4f90f768"} Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.690267 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:21 crc kubenswrapper[4989]: I1006 08:56:21.708871 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-cdb7w" podStartSLOduration=1.70885184 podStartE2EDuration="1.70885184s" podCreationTimestamp="2025-10-06 08:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:56:21.70718881 +0000 UTC m=+1032.497214440" watchObservedRunningTime="2025-10-06 08:56:21.70885184 +0000 UTC m=+1032.498877420" Oct 06 08:56:22 crc kubenswrapper[4989]: I1006 08:56:22.701800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sqqsn" event={"ID":"c021ef53-6c43-47fe-8429-f374626d4611","Type":"ContainerStarted","Data":"5feb123133585d41bde5bd856f6fc7d2332b772f22175d7ea70a0f255f84aa89"} Oct 06 08:56:22 crc kubenswrapper[4989]: I1006 08:56:22.727065 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-sqqsn" podStartSLOduration=2.727048718 podStartE2EDuration="2.727048718s" podCreationTimestamp="2025-10-06 08:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:56:22.724200382 +0000 UTC m=+1033.514225962" watchObservedRunningTime="2025-10-06 08:56:22.727048718 +0000 UTC m=+1033.517074298" Oct 06 08:56:23 crc kubenswrapper[4989]: I1006 08:56:23.708825 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-sqqsn" Oct 06 08:56:28 crc kubenswrapper[4989]: I1006 08:56:28.740895 4989 generic.go:334] "Generic (PLEG): container finished" podID="38c9054e-238b-4eef-8f71-ec00117efac9" containerID="d2c8ac5497cbe52be386f693bcf7a9a858aeb8dd46faca68b763aac12f1053ae" exitCode=0 Oct 06 08:56:28 crc kubenswrapper[4989]: I1006 08:56:28.740997 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerDied","Data":"d2c8ac5497cbe52be386f693bcf7a9a858aeb8dd46faca68b763aac12f1053ae"} Oct 06 08:56:28 crc kubenswrapper[4989]: I1006 08:56:28.742998 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" event={"ID":"a02ef7c3-577f-4284-88c6-ca16b3b335b8","Type":"ContainerStarted","Data":"087201f36338e318e4336e8f3c23f3ef4230978eb31872b6c795f082c528a1e1"} Oct 06 08:56:28 crc kubenswrapper[4989]: I1006 08:56:28.743347 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:28 crc kubenswrapper[4989]: I1006 08:56:28.797877 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" podStartSLOduration=1.694494937 podStartE2EDuration="8.797859895s" podCreationTimestamp="2025-10-06 08:56:20 +0000 UTC" firstStartedPulling="2025-10-06 08:56:20.697150475 +0000 UTC m=+1031.487176055" lastFinishedPulling="2025-10-06 08:56:27.800515433 +0000 UTC m=+1038.590541013" observedRunningTime="2025-10-06 08:56:28.792505871 +0000 UTC m=+1039.582531451" watchObservedRunningTime="2025-10-06 08:56:28.797859895 +0000 UTC m=+1039.587885475" Oct 06 08:56:29 crc kubenswrapper[4989]: I1006 08:56:29.750607 4989 generic.go:334] "Generic (PLEG): container finished" podID="38c9054e-238b-4eef-8f71-ec00117efac9" containerID="30d3f0e0283dee5344d9099f988d1a32fe4570f272f7c4e0f7b77a9884d7113a" exitCode=0 Oct 06 08:56:29 crc kubenswrapper[4989]: I1006 08:56:29.750727 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerDied","Data":"30d3f0e0283dee5344d9099f988d1a32fe4570f272f7c4e0f7b77a9884d7113a"} Oct 06 08:56:30 crc kubenswrapper[4989]: I1006 08:56:30.604533 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-cdb7w" Oct 06 08:56:30 crc kubenswrapper[4989]: I1006 08:56:30.757022 4989 generic.go:334] "Generic (PLEG): container finished" podID="38c9054e-238b-4eef-8f71-ec00117efac9" containerID="bf2ccfc1a2c304ea63038fd24a2957cc8ab1d53e53c28064342be62bd5b4f141" exitCode=0 Oct 06 08:56:30 crc kubenswrapper[4989]: I1006 08:56:30.757082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerDied","Data":"bf2ccfc1a2c304ea63038fd24a2957cc8ab1d53e53c28064342be62bd5b4f141"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.189445 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-sqqsn" Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769091 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"647da5b65f33268e8ce87b256678a2493c8480378cfe558a1a719ddbf423bf9f"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769143 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"d4f592a2dbf9d1bfc9438cd196c5c4092975ce48076ac7494aaf32c5894b3322"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769158 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"de64ca80bdc6dce9aa43e6508fa658875b3c2a728b7ecd89ead360ae67256957"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769172 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"6695df0f7c9409b4b0a93040437fbda180a7521100b19e74e7c89ce1836482c6"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769186 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"24dd1e8bf258a1813221407b4f4e3e9d7540a0ee772ad1ef97ab2e42e8fbd666"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769197 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gxdgx" event={"ID":"38c9054e-238b-4eef-8f71-ec00117efac9","Type":"ContainerStarted","Data":"25ddbfb355ada40738f61c319e650f78397bed5e4c195769e12b500c1aa0eddd"} Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.769300 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:31 crc kubenswrapper[4989]: I1006 08:56:31.825872 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-gxdgx" podStartSLOduration=4.672356317 podStartE2EDuration="11.825851519s" podCreationTimestamp="2025-10-06 08:56:20 +0000 UTC" firstStartedPulling="2025-10-06 08:56:20.654832327 +0000 UTC m=+1031.444857907" lastFinishedPulling="2025-10-06 08:56:27.808327529 +0000 UTC m=+1038.598353109" observedRunningTime="2025-10-06 08:56:31.822999226 +0000 UTC m=+1042.613024806" watchObservedRunningTime="2025-10-06 08:56:31.825851519 +0000 UTC m=+1042.615877099" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.680522 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c"] Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.681888 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.690358 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c"] Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.690647 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.729987 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.730096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.730133 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78j5x\" (UniqueName: \"kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.832354 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78j5x\" (UniqueName: \"kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.832503 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.832542 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.832979 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.833059 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:32 crc kubenswrapper[4989]: I1006 08:56:32.858058 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78j5x\" (UniqueName: \"kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.023135 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.241391 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c"] Oct 06 08:56:33 crc kubenswrapper[4989]: W1006 08:56:33.245970 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod180f1859_bedd_4e03_81f5_553e5bf46f2b.slice/crio-01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f WatchSource:0}: Error finding container 01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f: Status 404 returned error can't find the container with id 01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.781104 4989 generic.go:334] "Generic (PLEG): container finished" podID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerID="148324d4ad7e013e1721b96ff6b4ffbd69a8054c89f096b218533ecfb81a58ef" exitCode=0 Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.781158 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" event={"ID":"180f1859-bedd-4e03-81f5-553e5bf46f2b","Type":"ContainerDied","Data":"148324d4ad7e013e1721b96ff6b4ffbd69a8054c89f096b218533ecfb81a58ef"} Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.781513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" event={"ID":"180f1859-bedd-4e03-81f5-553e5bf46f2b","Type":"ContainerStarted","Data":"01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f"} Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.936200 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:56:33 crc kubenswrapper[4989]: I1006 08:56:33.936270 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:56:35 crc kubenswrapper[4989]: I1006 08:56:35.487171 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:35 crc kubenswrapper[4989]: I1006 08:56:35.528569 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:37 crc kubenswrapper[4989]: I1006 08:56:37.806143 4989 generic.go:334] "Generic (PLEG): container finished" podID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerID="d692a1c262610418e2aa7193cba9bbe8a50d9c36c6e3acc0b944c8412016a47e" exitCode=0 Oct 06 08:56:37 crc kubenswrapper[4989]: I1006 08:56:37.806220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" event={"ID":"180f1859-bedd-4e03-81f5-553e5bf46f2b","Type":"ContainerDied","Data":"d692a1c262610418e2aa7193cba9bbe8a50d9c36c6e3acc0b944c8412016a47e"} Oct 06 08:56:38 crc kubenswrapper[4989]: I1006 08:56:38.816771 4989 generic.go:334] "Generic (PLEG): container finished" podID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerID="2f074321112495b553e581863d94ad5b8c52b197c3c35d2ec7595f1a67d2a347" exitCode=0 Oct 06 08:56:38 crc kubenswrapper[4989]: I1006 08:56:38.816835 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" event={"ID":"180f1859-bedd-4e03-81f5-553e5bf46f2b","Type":"ContainerDied","Data":"2f074321112495b553e581863d94ad5b8c52b197c3c35d2ec7595f1a67d2a347"} Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.074535 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.234392 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle\") pod \"180f1859-bedd-4e03-81f5-553e5bf46f2b\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.234435 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78j5x\" (UniqueName: \"kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x\") pod \"180f1859-bedd-4e03-81f5-553e5bf46f2b\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.234520 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util\") pod \"180f1859-bedd-4e03-81f5-553e5bf46f2b\" (UID: \"180f1859-bedd-4e03-81f5-553e5bf46f2b\") " Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.235615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle" (OuterVolumeSpecName: "bundle") pod "180f1859-bedd-4e03-81f5-553e5bf46f2b" (UID: "180f1859-bedd-4e03-81f5-553e5bf46f2b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.241891 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x" (OuterVolumeSpecName: "kube-api-access-78j5x") pod "180f1859-bedd-4e03-81f5-553e5bf46f2b" (UID: "180f1859-bedd-4e03-81f5-553e5bf46f2b"). InnerVolumeSpecName "kube-api-access-78j5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.246792 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util" (OuterVolumeSpecName: "util") pod "180f1859-bedd-4e03-81f5-553e5bf46f2b" (UID: "180f1859-bedd-4e03-81f5-553e5bf46f2b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.337811 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-util\") on node \"crc\" DevicePath \"\"" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.338033 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/180f1859-bedd-4e03-81f5-553e5bf46f2b-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.338044 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78j5x\" (UniqueName: \"kubernetes.io/projected/180f1859-bedd-4e03-81f5-553e5bf46f2b-kube-api-access-78j5x\") on node \"crc\" DevicePath \"\"" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.489221 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-gxdgx" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.499793 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-bs986" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.833711 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" event={"ID":"180f1859-bedd-4e03-81f5-553e5bf46f2b","Type":"ContainerDied","Data":"01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f"} Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.833756 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01f06342786c08d5e58f141cff69eabca58abe6baf333b4b1936bb8a70e47e9f" Oct 06 08:56:40 crc kubenswrapper[4989]: I1006 08:56:40.833836 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.124587 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw"] Oct 06 08:56:58 crc kubenswrapper[4989]: E1006 08:56:58.125313 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="extract" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.125326 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="extract" Oct 06 08:56:58 crc kubenswrapper[4989]: E1006 08:56:58.125337 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="util" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.125344 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="util" Oct 06 08:56:58 crc kubenswrapper[4989]: E1006 08:56:58.125367 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="pull" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.125373 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="pull" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.125474 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="180f1859-bedd-4e03-81f5-553e5bf46f2b" containerName="extract" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.125905 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.128033 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.130206 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.131199 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-pxrf2" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.138306 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw"] Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.164237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br5zl\" (UniqueName: \"kubernetes.io/projected/0e79a09d-66e2-4275-86a0-1a3db06141d2-kube-api-access-br5zl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-flvgw\" (UID: \"0e79a09d-66e2-4275-86a0-1a3db06141d2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.265421 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br5zl\" (UniqueName: \"kubernetes.io/projected/0e79a09d-66e2-4275-86a0-1a3db06141d2-kube-api-access-br5zl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-flvgw\" (UID: \"0e79a09d-66e2-4275-86a0-1a3db06141d2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.295749 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br5zl\" (UniqueName: \"kubernetes.io/projected/0e79a09d-66e2-4275-86a0-1a3db06141d2-kube-api-access-br5zl\") pod \"cert-manager-operator-controller-manager-57cd46d6d-flvgw\" (UID: \"0e79a09d-66e2-4275-86a0-1a3db06141d2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.442600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" Oct 06 08:56:58 crc kubenswrapper[4989]: I1006 08:56:58.988028 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw"] Oct 06 08:56:58 crc kubenswrapper[4989]: W1006 08:56:58.991683 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e79a09d_66e2_4275_86a0_1a3db06141d2.slice/crio-dfd9cd12915f6b6786c9439ba156101f162d7614a59ccbd3fbfa08f87d4e5565 WatchSource:0}: Error finding container dfd9cd12915f6b6786c9439ba156101f162d7614a59ccbd3fbfa08f87d4e5565: Status 404 returned error can't find the container with id dfd9cd12915f6b6786c9439ba156101f162d7614a59ccbd3fbfa08f87d4e5565 Oct 06 08:56:59 crc kubenswrapper[4989]: I1006 08:56:59.951768 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" event={"ID":"0e79a09d-66e2-4275-86a0-1a3db06141d2","Type":"ContainerStarted","Data":"dfd9cd12915f6b6786c9439ba156101f162d7614a59ccbd3fbfa08f87d4e5565"} Oct 06 08:57:03 crc kubenswrapper[4989]: I1006 08:57:03.937148 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:57:03 crc kubenswrapper[4989]: I1006 08:57:03.937799 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:57:03 crc kubenswrapper[4989]: I1006 08:57:03.944893 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 08:57:03 crc kubenswrapper[4989]: I1006 08:57:03.945993 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 08:57:03 crc kubenswrapper[4989]: I1006 08:57:03.946053 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f" gracePeriod=600 Oct 06 08:57:04 crc kubenswrapper[4989]: I1006 08:57:04.989164 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f" exitCode=0 Oct 06 08:57:04 crc kubenswrapper[4989]: I1006 08:57:04.989206 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f"} Oct 06 08:57:04 crc kubenswrapper[4989]: I1006 08:57:04.989244 4989 scope.go:117] "RemoveContainer" containerID="b69d6ad6b29f029a03841d0631e19025c62893f3539f8fd296156523a92a190a" Oct 06 08:57:07 crc kubenswrapper[4989]: I1006 08:57:07.002876 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" event={"ID":"0e79a09d-66e2-4275-86a0-1a3db06141d2","Type":"ContainerStarted","Data":"fc8e955b540a262f4b28c4c9289137d079743a47852bfe4f95b25515bd5199a0"} Oct 06 08:57:07 crc kubenswrapper[4989]: I1006 08:57:07.006238 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050"} Oct 06 08:57:07 crc kubenswrapper[4989]: I1006 08:57:07.023222 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-flvgw" podStartSLOduration=1.562409808 podStartE2EDuration="9.023206074s" podCreationTimestamp="2025-10-06 08:56:58 +0000 UTC" firstStartedPulling="2025-10-06 08:56:58.994089415 +0000 UTC m=+1069.784114995" lastFinishedPulling="2025-10-06 08:57:06.454885681 +0000 UTC m=+1077.244911261" observedRunningTime="2025-10-06 08:57:07.021819573 +0000 UTC m=+1077.811845163" watchObservedRunningTime="2025-10-06 08:57:07.023206074 +0000 UTC m=+1077.813231654" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.674422 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-9djb6"] Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.675735 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.677843 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.678047 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.678209 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-sdn8n" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.683438 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-9djb6"] Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.730032 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lhrq\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-kube-api-access-6lhrq\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.730126 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-bound-sa-token\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.830857 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lhrq\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-kube-api-access-6lhrq\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.830984 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-bound-sa-token\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.849419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lhrq\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-kube-api-access-6lhrq\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.853368 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/48613a5d-3bd1-4731-ab29-efe554110a70-bound-sa-token\") pod \"cert-manager-webhook-d969966f-9djb6\" (UID: \"48613a5d-3bd1-4731-ab29-efe554110a70\") " pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:10 crc kubenswrapper[4989]: I1006 08:57:10.997643 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.288195 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh"] Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.289178 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.291431 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-9sbqw" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.301154 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh"] Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.330144 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-9djb6"] Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.343064 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.439614 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr5g4\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-kube-api-access-pr5g4\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.439711 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.541076 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr5g4\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-kube-api-access-pr5g4\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.541397 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.558798 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr5g4\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-kube-api-access-pr5g4\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.558803 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6be97c71-f240-44ab-bba5-f52434740520-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-5d6bh\" (UID: \"6be97c71-f240-44ab-bba5-f52434740520\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.611096 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" Oct 06 08:57:11 crc kubenswrapper[4989]: I1006 08:57:11.814804 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh"] Oct 06 08:57:12 crc kubenswrapper[4989]: I1006 08:57:12.031682 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" event={"ID":"6be97c71-f240-44ab-bba5-f52434740520","Type":"ContainerStarted","Data":"6f641f448152289b5b09584e245f67836fe3a68fd6abda8348ea0c9c0861c4ff"} Oct 06 08:57:12 crc kubenswrapper[4989]: I1006 08:57:12.032560 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" event={"ID":"48613a5d-3bd1-4731-ab29-efe554110a70","Type":"ContainerStarted","Data":"a100de9e005b91f98a8350daaeb51edba93219ee2d3ba32bab208f95d6c94cd1"} Oct 06 08:57:17 crc kubenswrapper[4989]: I1006 08:57:17.083555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" event={"ID":"48613a5d-3bd1-4731-ab29-efe554110a70","Type":"ContainerStarted","Data":"eb8ff716ea7b4cbe8561f6522267d36f068e919ad85584a880125d6e443d147d"} Oct 06 08:57:17 crc kubenswrapper[4989]: I1006 08:57:17.084139 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:17 crc kubenswrapper[4989]: I1006 08:57:17.085345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" event={"ID":"6be97c71-f240-44ab-bba5-f52434740520","Type":"ContainerStarted","Data":"66a132008b9874d16a37bf750ee60dfe849b8e5db69bd542010d554439d7e4dc"} Oct 06 08:57:17 crc kubenswrapper[4989]: I1006 08:57:17.103862 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" podStartSLOduration=2.088828912 podStartE2EDuration="7.103832366s" podCreationTimestamp="2025-10-06 08:57:10 +0000 UTC" firstStartedPulling="2025-10-06 08:57:11.342783877 +0000 UTC m=+1082.132809457" lastFinishedPulling="2025-10-06 08:57:16.357787321 +0000 UTC m=+1087.147812911" observedRunningTime="2025-10-06 08:57:17.099641755 +0000 UTC m=+1087.889667345" watchObservedRunningTime="2025-10-06 08:57:17.103832366 +0000 UTC m=+1087.893857986" Oct 06 08:57:17 crc kubenswrapper[4989]: I1006 08:57:17.116265 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-5d6bh" podStartSLOduration=1.571315126 podStartE2EDuration="6.116216524s" podCreationTimestamp="2025-10-06 08:57:11 +0000 UTC" firstStartedPulling="2025-10-06 08:57:11.830512242 +0000 UTC m=+1082.620537822" lastFinishedPulling="2025-10-06 08:57:16.37541364 +0000 UTC m=+1087.165439220" observedRunningTime="2025-10-06 08:57:17.113127865 +0000 UTC m=+1087.903153445" watchObservedRunningTime="2025-10-06 08:57:17.116216524 +0000 UTC m=+1087.906242124" Oct 06 08:57:26 crc kubenswrapper[4989]: I1006 08:57:26.001073 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-9djb6" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.420205 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-pzq52"] Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.422253 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.426173 4989 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b9nhk" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.440212 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-pzq52"] Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.574646 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrt8q\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-kube-api-access-nrt8q\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.574717 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.676545 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrt8q\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-kube-api-access-nrt8q\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.676634 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.698719 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrt8q\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-kube-api-access-nrt8q\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.699381 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f0a0cf7b-ef3f-4b08-a574-9f81f13d227b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-pzq52\" (UID: \"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b\") " pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:29 crc kubenswrapper[4989]: I1006 08:57:29.757554 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" Oct 06 08:57:30 crc kubenswrapper[4989]: I1006 08:57:30.194751 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-pzq52"] Oct 06 08:57:30 crc kubenswrapper[4989]: W1006 08:57:30.206573 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0a0cf7b_ef3f_4b08_a574_9f81f13d227b.slice/crio-b4e06cec071f22dfc52b997255a6a5c4b733a98f84f2b400c274c95e88e7b9e5 WatchSource:0}: Error finding container b4e06cec071f22dfc52b997255a6a5c4b733a98f84f2b400c274c95e88e7b9e5: Status 404 returned error can't find the container with id b4e06cec071f22dfc52b997255a6a5c4b733a98f84f2b400c274c95e88e7b9e5 Oct 06 08:57:31 crc kubenswrapper[4989]: I1006 08:57:31.181262 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" event={"ID":"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b","Type":"ContainerStarted","Data":"ac63a3bcc8aa1e457cd02793cbf0ab03b27957e06e765f33cbe219e2d1096572"} Oct 06 08:57:31 crc kubenswrapper[4989]: I1006 08:57:31.181632 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" event={"ID":"f0a0cf7b-ef3f-4b08-a574-9f81f13d227b","Type":"ContainerStarted","Data":"b4e06cec071f22dfc52b997255a6a5c4b733a98f84f2b400c274c95e88e7b9e5"} Oct 06 08:57:31 crc kubenswrapper[4989]: I1006 08:57:31.209409 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-pzq52" podStartSLOduration=2.209382989 podStartE2EDuration="2.209382989s" podCreationTimestamp="2025-10-06 08:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:57:31.20387868 +0000 UTC m=+1101.993904300" watchObservedRunningTime="2025-10-06 08:57:31.209382989 +0000 UTC m=+1101.999408609" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.618066 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.619671 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.628901 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.628981 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.629068 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-fc4cp" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.640700 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.720920 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jctcm\" (UniqueName: \"kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm\") pod \"openstack-operator-index-29hrw\" (UID: \"54c0b97b-7d8c-43d4-9709-36226e2525af\") " pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.822843 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jctcm\" (UniqueName: \"kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm\") pod \"openstack-operator-index-29hrw\" (UID: \"54c0b97b-7d8c-43d4-9709-36226e2525af\") " pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.842430 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jctcm\" (UniqueName: \"kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm\") pod \"openstack-operator-index-29hrw\" (UID: \"54c0b97b-7d8c-43d4-9709-36226e2525af\") " pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:39 crc kubenswrapper[4989]: I1006 08:57:39.938670 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:40 crc kubenswrapper[4989]: I1006 08:57:40.355563 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:40 crc kubenswrapper[4989]: W1006 08:57:40.366290 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54c0b97b_7d8c_43d4_9709_36226e2525af.slice/crio-5a1d3098b784a87fd7fb7cec9641b91de7e6f3a9012c815a547c3abdd5174ec6 WatchSource:0}: Error finding container 5a1d3098b784a87fd7fb7cec9641b91de7e6f3a9012c815a547c3abdd5174ec6: Status 404 returned error can't find the container with id 5a1d3098b784a87fd7fb7cec9641b91de7e6f3a9012c815a547c3abdd5174ec6 Oct 06 08:57:41 crc kubenswrapper[4989]: I1006 08:57:41.250634 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29hrw" event={"ID":"54c0b97b-7d8c-43d4-9709-36226e2525af","Type":"ContainerStarted","Data":"5a1d3098b784a87fd7fb7cec9641b91de7e6f3a9012c815a547c3abdd5174ec6"} Oct 06 08:57:42 crc kubenswrapper[4989]: I1006 08:57:42.983662 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.265763 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29hrw" event={"ID":"54c0b97b-7d8c-43d4-9709-36226e2525af","Type":"ContainerStarted","Data":"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe"} Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.280273 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-29hrw" podStartSLOduration=2.466994345 podStartE2EDuration="4.280258076s" podCreationTimestamp="2025-10-06 08:57:39 +0000 UTC" firstStartedPulling="2025-10-06 08:57:40.368037447 +0000 UTC m=+1111.158063027" lastFinishedPulling="2025-10-06 08:57:42.181301168 +0000 UTC m=+1112.971326758" observedRunningTime="2025-10-06 08:57:43.278610198 +0000 UTC m=+1114.068635778" watchObservedRunningTime="2025-10-06 08:57:43.280258076 +0000 UTC m=+1114.070283656" Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.583186 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jjkjt"] Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.584141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.592972 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjkjt"] Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.684920 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxzst\" (UniqueName: \"kubernetes.io/projected/3c87cc47-079e-44f3-bac3-a66ce98b1692-kube-api-access-pxzst\") pod \"openstack-operator-index-jjkjt\" (UID: \"3c87cc47-079e-44f3-bac3-a66ce98b1692\") " pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.786636 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxzst\" (UniqueName: \"kubernetes.io/projected/3c87cc47-079e-44f3-bac3-a66ce98b1692-kube-api-access-pxzst\") pod \"openstack-operator-index-jjkjt\" (UID: \"3c87cc47-079e-44f3-bac3-a66ce98b1692\") " pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.804720 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxzst\" (UniqueName: \"kubernetes.io/projected/3c87cc47-079e-44f3-bac3-a66ce98b1692-kube-api-access-pxzst\") pod \"openstack-operator-index-jjkjt\" (UID: \"3c87cc47-079e-44f3-bac3-a66ce98b1692\") " pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:43 crc kubenswrapper[4989]: I1006 08:57:43.903371 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.270375 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-29hrw" podUID="54c0b97b-7d8c-43d4-9709-36226e2525af" containerName="registry-server" containerID="cri-o://af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe" gracePeriod=2 Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.308392 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjkjt"] Oct 06 08:57:44 crc kubenswrapper[4989]: W1006 08:57:44.358284 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c87cc47_079e_44f3_bac3_a66ce98b1692.slice/crio-5514458d769e0346390d12170b674768534387c8c0f606db6c46a6151083f64f WatchSource:0}: Error finding container 5514458d769e0346390d12170b674768534387c8c0f606db6c46a6151083f64f: Status 404 returned error can't find the container with id 5514458d769e0346390d12170b674768534387c8c0f606db6c46a6151083f64f Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.576759 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.696421 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jctcm\" (UniqueName: \"kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm\") pod \"54c0b97b-7d8c-43d4-9709-36226e2525af\" (UID: \"54c0b97b-7d8c-43d4-9709-36226e2525af\") " Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.702915 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm" (OuterVolumeSpecName: "kube-api-access-jctcm") pod "54c0b97b-7d8c-43d4-9709-36226e2525af" (UID: "54c0b97b-7d8c-43d4-9709-36226e2525af"). InnerVolumeSpecName "kube-api-access-jctcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:57:44 crc kubenswrapper[4989]: I1006 08:57:44.798413 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jctcm\" (UniqueName: \"kubernetes.io/projected/54c0b97b-7d8c-43d4-9709-36226e2525af-kube-api-access-jctcm\") on node \"crc\" DevicePath \"\"" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.277454 4989 generic.go:334] "Generic (PLEG): container finished" podID="54c0b97b-7d8c-43d4-9709-36226e2525af" containerID="af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe" exitCode=0 Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.277549 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-29hrw" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.277557 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29hrw" event={"ID":"54c0b97b-7d8c-43d4-9709-36226e2525af","Type":"ContainerDied","Data":"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe"} Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.277701 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-29hrw" event={"ID":"54c0b97b-7d8c-43d4-9709-36226e2525af","Type":"ContainerDied","Data":"5a1d3098b784a87fd7fb7cec9641b91de7e6f3a9012c815a547c3abdd5174ec6"} Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.277728 4989 scope.go:117] "RemoveContainer" containerID="af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.279449 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjkjt" event={"ID":"3c87cc47-079e-44f3-bac3-a66ce98b1692","Type":"ContainerStarted","Data":"905ceac1ff9dab4d2af78b7a598ab52d4ba91347ff51e658a97d413ab26eaa7f"} Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.279485 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjkjt" event={"ID":"3c87cc47-079e-44f3-bac3-a66ce98b1692","Type":"ContainerStarted","Data":"5514458d769e0346390d12170b674768534387c8c0f606db6c46a6151083f64f"} Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.295471 4989 scope.go:117] "RemoveContainer" containerID="af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe" Oct 06 08:57:45 crc kubenswrapper[4989]: E1006 08:57:45.296014 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe\": container with ID starting with af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe not found: ID does not exist" containerID="af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.296044 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe"} err="failed to get container status \"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe\": rpc error: code = NotFound desc = could not find container \"af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe\": container with ID starting with af4459380e3ae2fc1b1de7f6511d3eeaa0c3309e64e785db51901842ec4e1dbe not found: ID does not exist" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.299465 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jjkjt" podStartSLOduration=2.2565898300000002 podStartE2EDuration="2.299451066s" podCreationTimestamp="2025-10-06 08:57:43 +0000 UTC" firstStartedPulling="2025-10-06 08:57:44.364020835 +0000 UTC m=+1115.154046445" lastFinishedPulling="2025-10-06 08:57:44.406882101 +0000 UTC m=+1115.196907681" observedRunningTime="2025-10-06 08:57:45.297523211 +0000 UTC m=+1116.087548791" watchObservedRunningTime="2025-10-06 08:57:45.299451066 +0000 UTC m=+1116.089476636" Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.312477 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.316595 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-29hrw"] Oct 06 08:57:45 crc kubenswrapper[4989]: I1006 08:57:45.945537 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c0b97b-7d8c-43d4-9709-36226e2525af" path="/var/lib/kubelet/pods/54c0b97b-7d8c-43d4-9709-36226e2525af/volumes" Oct 06 08:57:53 crc kubenswrapper[4989]: I1006 08:57:53.904413 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:53 crc kubenswrapper[4989]: I1006 08:57:53.904754 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:53 crc kubenswrapper[4989]: I1006 08:57:53.944680 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:54 crc kubenswrapper[4989]: I1006 08:57:54.359828 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-jjkjt" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.629451 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd"] Oct 06 08:57:59 crc kubenswrapper[4989]: E1006 08:57:59.630154 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c0b97b-7d8c-43d4-9709-36226e2525af" containerName="registry-server" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.630169 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c0b97b-7d8c-43d4-9709-36226e2525af" containerName="registry-server" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.630300 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c0b97b-7d8c-43d4-9709-36226e2525af" containerName="registry-server" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.631398 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.633384 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cvbn7" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.648690 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd"] Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.710127 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m4gf\" (UniqueName: \"kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.710299 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.710427 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.812666 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.812738 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.812799 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m4gf\" (UniqueName: \"kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.813675 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.814049 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.833269 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m4gf\" (UniqueName: \"kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf\") pod \"cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:57:59 crc kubenswrapper[4989]: I1006 08:57:59.990504 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:58:00 crc kubenswrapper[4989]: I1006 08:58:00.503761 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd"] Oct 06 08:58:01 crc kubenswrapper[4989]: I1006 08:58:01.371942 4989 generic.go:334] "Generic (PLEG): container finished" podID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerID="8e952b5575f904e1af8972fada3e6c83d2720c5cc4aa3ebde1e4b246abeb176d" exitCode=0 Oct 06 08:58:01 crc kubenswrapper[4989]: I1006 08:58:01.372022 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" event={"ID":"ba2e51bd-d38a-41e4-91eb-853d18b5a459","Type":"ContainerDied","Data":"8e952b5575f904e1af8972fada3e6c83d2720c5cc4aa3ebde1e4b246abeb176d"} Oct 06 08:58:01 crc kubenswrapper[4989]: I1006 08:58:01.372267 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" event={"ID":"ba2e51bd-d38a-41e4-91eb-853d18b5a459","Type":"ContainerStarted","Data":"eb5a9d690767666cb013fa31e9b39fccc7c2081027d8778be8c7d0224af145fd"} Oct 06 08:58:03 crc kubenswrapper[4989]: I1006 08:58:03.388218 4989 generic.go:334] "Generic (PLEG): container finished" podID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerID="c6abe9819c67d13dcbb3ea3acfac56b276c5af9a511085e4f99b6e8a16a0e46d" exitCode=0 Oct 06 08:58:03 crc kubenswrapper[4989]: I1006 08:58:03.388409 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" event={"ID":"ba2e51bd-d38a-41e4-91eb-853d18b5a459","Type":"ContainerDied","Data":"c6abe9819c67d13dcbb3ea3acfac56b276c5af9a511085e4f99b6e8a16a0e46d"} Oct 06 08:58:04 crc kubenswrapper[4989]: I1006 08:58:04.397963 4989 generic.go:334] "Generic (PLEG): container finished" podID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerID="29d50e691d92fd4c2311cec0d284300e56bdfa8e1f448f6b2288a80616bd97d9" exitCode=0 Oct 06 08:58:04 crc kubenswrapper[4989]: I1006 08:58:04.398077 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" event={"ID":"ba2e51bd-d38a-41e4-91eb-853d18b5a459","Type":"ContainerDied","Data":"29d50e691d92fd4c2311cec0d284300e56bdfa8e1f448f6b2288a80616bd97d9"} Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.651140 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.700003 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle\") pod \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.700114 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m4gf\" (UniqueName: \"kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf\") pod \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.700142 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util\") pod \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\" (UID: \"ba2e51bd-d38a-41e4-91eb-853d18b5a459\") " Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.701288 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle" (OuterVolumeSpecName: "bundle") pod "ba2e51bd-d38a-41e4-91eb-853d18b5a459" (UID: "ba2e51bd-d38a-41e4-91eb-853d18b5a459"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.706755 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf" (OuterVolumeSpecName: "kube-api-access-4m4gf") pod "ba2e51bd-d38a-41e4-91eb-853d18b5a459" (UID: "ba2e51bd-d38a-41e4-91eb-853d18b5a459"). InnerVolumeSpecName "kube-api-access-4m4gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.713412 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util" (OuterVolumeSpecName: "util") pod "ba2e51bd-d38a-41e4-91eb-853d18b5a459" (UID: "ba2e51bd-d38a-41e4-91eb-853d18b5a459"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.801787 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.801822 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m4gf\" (UniqueName: \"kubernetes.io/projected/ba2e51bd-d38a-41e4-91eb-853d18b5a459-kube-api-access-4m4gf\") on node \"crc\" DevicePath \"\"" Oct 06 08:58:05 crc kubenswrapper[4989]: I1006 08:58:05.801834 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ba2e51bd-d38a-41e4-91eb-853d18b5a459-util\") on node \"crc\" DevicePath \"\"" Oct 06 08:58:06 crc kubenswrapper[4989]: I1006 08:58:06.417552 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" event={"ID":"ba2e51bd-d38a-41e4-91eb-853d18b5a459","Type":"ContainerDied","Data":"eb5a9d690767666cb013fa31e9b39fccc7c2081027d8778be8c7d0224af145fd"} Oct 06 08:58:06 crc kubenswrapper[4989]: I1006 08:58:06.418199 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb5a9d690767666cb013fa31e9b39fccc7c2081027d8778be8c7d0224af145fd" Oct 06 08:58:06 crc kubenswrapper[4989]: I1006 08:58:06.417637 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.197378 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt"] Oct 06 08:58:12 crc kubenswrapper[4989]: E1006 08:58:12.198167 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="extract" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.198181 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="extract" Oct 06 08:58:12 crc kubenswrapper[4989]: E1006 08:58:12.198199 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="pull" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.198206 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="pull" Oct 06 08:58:12 crc kubenswrapper[4989]: E1006 08:58:12.198220 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="util" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.198227 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="util" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.198339 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba2e51bd-d38a-41e4-91eb-853d18b5a459" containerName="extract" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.199040 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.201205 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-z7h47" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.294430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7vj7\" (UniqueName: \"kubernetes.io/projected/2b3e23ba-5e26-49ef-acaf-0520111b142a-kube-api-access-v7vj7\") pod \"openstack-operator-controller-operator-6497dff45c-mpcvt\" (UID: \"2b3e23ba-5e26-49ef-acaf-0520111b142a\") " pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.325570 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt"] Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.395845 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7vj7\" (UniqueName: \"kubernetes.io/projected/2b3e23ba-5e26-49ef-acaf-0520111b142a-kube-api-access-v7vj7\") pod \"openstack-operator-controller-operator-6497dff45c-mpcvt\" (UID: \"2b3e23ba-5e26-49ef-acaf-0520111b142a\") " pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.414202 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7vj7\" (UniqueName: \"kubernetes.io/projected/2b3e23ba-5e26-49ef-acaf-0520111b142a-kube-api-access-v7vj7\") pod \"openstack-operator-controller-operator-6497dff45c-mpcvt\" (UID: \"2b3e23ba-5e26-49ef-acaf-0520111b142a\") " pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.514433 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:12 crc kubenswrapper[4989]: I1006 08:58:12.777301 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt"] Oct 06 08:58:12 crc kubenswrapper[4989]: W1006 08:58:12.779723 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b3e23ba_5e26_49ef_acaf_0520111b142a.slice/crio-36d05a925bbc33bd41fdd1a968eee06456e056746ac1f39c59c98544b5f410aa WatchSource:0}: Error finding container 36d05a925bbc33bd41fdd1a968eee06456e056746ac1f39c59c98544b5f410aa: Status 404 returned error can't find the container with id 36d05a925bbc33bd41fdd1a968eee06456e056746ac1f39c59c98544b5f410aa Oct 06 08:58:13 crc kubenswrapper[4989]: I1006 08:58:13.469292 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" event={"ID":"2b3e23ba-5e26-49ef-acaf-0520111b142a","Type":"ContainerStarted","Data":"36d05a925bbc33bd41fdd1a968eee06456e056746ac1f39c59c98544b5f410aa"} Oct 06 08:58:17 crc kubenswrapper[4989]: I1006 08:58:17.493352 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" event={"ID":"2b3e23ba-5e26-49ef-acaf-0520111b142a","Type":"ContainerStarted","Data":"8993ae020a1c69a71ed46b4afd86b6b1238d02ff05ca5890a766db5b584cdada"} Oct 06 08:58:19 crc kubenswrapper[4989]: I1006 08:58:19.506715 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" event={"ID":"2b3e23ba-5e26-49ef-acaf-0520111b142a","Type":"ContainerStarted","Data":"0decd35cb59d84409b959ddb60ee098934a5e2ec0cf3a27165b9cb43512ac17f"} Oct 06 08:58:19 crc kubenswrapper[4989]: I1006 08:58:19.507082 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:22 crc kubenswrapper[4989]: I1006 08:58:22.517752 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" Oct 06 08:58:22 crc kubenswrapper[4989]: I1006 08:58:22.554023 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6497dff45c-mpcvt" podStartSLOduration=4.326041715 podStartE2EDuration="10.554005421s" podCreationTimestamp="2025-10-06 08:58:12 +0000 UTC" firstStartedPulling="2025-10-06 08:58:12.782322391 +0000 UTC m=+1143.572347961" lastFinishedPulling="2025-10-06 08:58:19.010286087 +0000 UTC m=+1149.800311667" observedRunningTime="2025-10-06 08:58:19.544480125 +0000 UTC m=+1150.334505715" watchObservedRunningTime="2025-10-06 08:58:22.554005421 +0000 UTC m=+1153.344031001" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.154727 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.156055 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.159151 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-knbr7" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.175361 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.176304 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.178119 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lbrk4" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.181157 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.187545 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.188812 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.191225 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hvtqb" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.196409 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.218677 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.233707 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.234663 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.240120 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-b4kxm" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.240892 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfg9t\" (UniqueName: \"kubernetes.io/projected/c0f513bb-5736-42bd-a6cb-cf5314b940ee-kube-api-access-qfg9t\") pod \"designate-operator-controller-manager-75dfd9b554-8h5sx\" (UID: \"c0f513bb-5736-42bd-a6cb-cf5314b940ee\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.240958 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2hmh\" (UniqueName: \"kubernetes.io/projected/9e971444-f7c2-4443-ad67-8344cc2ece01-kube-api-access-t2hmh\") pod \"cinder-operator-controller-manager-7d4d4f8d-6r4p2\" (UID: \"9e971444-f7c2-4443-ad67-8344cc2ece01\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.240993 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hktvj\" (UniqueName: \"kubernetes.io/projected/9ed264b3-7650-41ed-bb45-c5e0265ebcaf-kube-api-access-hktvj\") pod \"barbican-operator-controller-manager-5f7c849b98-4skd5\" (UID: \"9ed264b3-7650-41ed-bb45-c5e0265ebcaf\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.249869 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.284981 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.287453 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.307857 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-hdk7d" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.324393 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.325493 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.330984 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5krlm" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.342589 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.343538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2hmh\" (UniqueName: \"kubernetes.io/projected/9e971444-f7c2-4443-ad67-8344cc2ece01-kube-api-access-t2hmh\") pod \"cinder-operator-controller-manager-7d4d4f8d-6r4p2\" (UID: \"9e971444-f7c2-4443-ad67-8344cc2ece01\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.343722 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hktvj\" (UniqueName: \"kubernetes.io/projected/9ed264b3-7650-41ed-bb45-c5e0265ebcaf-kube-api-access-hktvj\") pod \"barbican-operator-controller-manager-5f7c849b98-4skd5\" (UID: \"9ed264b3-7650-41ed-bb45-c5e0265ebcaf\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.343875 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqcgf\" (UniqueName: \"kubernetes.io/projected/cce21119-33b5-4f7c-834c-0e8f16ccd825-kube-api-access-mqcgf\") pod \"glance-operator-controller-manager-5568b5d68-25cxf\" (UID: \"cce21119-33b5-4f7c-834c-0e8f16ccd825\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.344001 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq56r\" (UniqueName: \"kubernetes.io/projected/30e9d8ad-e772-44e4-8234-a0ef8a3483e1-kube-api-access-mq56r\") pod \"heat-operator-controller-manager-8f58bc9db-phj9c\" (UID: \"30e9d8ad-e772-44e4-8234-a0ef8a3483e1\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.344147 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfg9t\" (UniqueName: \"kubernetes.io/projected/c0f513bb-5736-42bd-a6cb-cf5314b940ee-kube-api-access-qfg9t\") pod \"designate-operator-controller-manager-75dfd9b554-8h5sx\" (UID: \"c0f513bb-5736-42bd-a6cb-cf5314b940ee\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.367930 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.386783 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.388072 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.388902 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2hmh\" (UniqueName: \"kubernetes.io/projected/9e971444-f7c2-4443-ad67-8344cc2ece01-kube-api-access-t2hmh\") pod \"cinder-operator-controller-manager-7d4d4f8d-6r4p2\" (UID: \"9e971444-f7c2-4443-ad67-8344cc2ece01\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.392513 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfg9t\" (UniqueName: \"kubernetes.io/projected/c0f513bb-5736-42bd-a6cb-cf5314b940ee-kube-api-access-qfg9t\") pod \"designate-operator-controller-manager-75dfd9b554-8h5sx\" (UID: \"c0f513bb-5736-42bd-a6cb-cf5314b940ee\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.395303 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.396890 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.397025 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-c9jks" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.397975 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.403875 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-n6dcj" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.405836 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.408232 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hktvj\" (UniqueName: \"kubernetes.io/projected/9ed264b3-7650-41ed-bb45-c5e0265ebcaf-kube-api-access-hktvj\") pod \"barbican-operator-controller-manager-5f7c849b98-4skd5\" (UID: \"9ed264b3-7650-41ed-bb45-c5e0265ebcaf\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.413822 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.414911 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.417146 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2zlwl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.418534 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.428591 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.436962 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.437957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.443139 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nzb7b" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqcgf\" (UniqueName: \"kubernetes.io/projected/cce21119-33b5-4f7c-834c-0e8f16ccd825-kube-api-access-mqcgf\") pod \"glance-operator-controller-manager-5568b5d68-25cxf\" (UID: \"cce21119-33b5-4f7c-834c-0e8f16ccd825\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447430 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq56r\" (UniqueName: \"kubernetes.io/projected/30e9d8ad-e772-44e4-8234-a0ef8a3483e1-kube-api-access-mq56r\") pod \"heat-operator-controller-manager-8f58bc9db-phj9c\" (UID: \"30e9d8ad-e772-44e4-8234-a0ef8a3483e1\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447476 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447492 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k2t6\" (UniqueName: \"kubernetes.io/projected/b1667c4d-4d3a-41b1-a02a-890a2b0f26ea-kube-api-access-7k2t6\") pod \"horizon-operator-controller-manager-54876c876f-dqjpg\" (UID: \"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447520 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pgv4\" (UniqueName: \"kubernetes.io/projected/05b084da-1f15-4dba-8d4e-b540d4447318-kube-api-access-2pgv4\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c76wv\" (UniqueName: \"kubernetes.io/projected/45344ba0-bd44-4a54-9628-460586eae8df-kube-api-access-c76wv\") pod \"keystone-operator-controller-manager-655d88ccb9-79ksl\" (UID: \"45344ba0-bd44-4a54-9628-460586eae8df\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.447567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr8q4\" (UniqueName: \"kubernetes.io/projected/b89d614b-bb9f-4ba5-bb16-d9fe9731279f-kube-api-access-rr8q4\") pod \"ironic-operator-controller-manager-699b87f775-mrz27\" (UID: \"b89d614b-bb9f-4ba5-bb16-d9fe9731279f\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.452606 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.458038 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.459300 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.464018 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-fs2zv" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.470481 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.471360 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq56r\" (UniqueName: \"kubernetes.io/projected/30e9d8ad-e772-44e4-8234-a0ef8a3483e1-kube-api-access-mq56r\") pod \"heat-operator-controller-manager-8f58bc9db-phj9c\" (UID: \"30e9d8ad-e772-44e4-8234-a0ef8a3483e1\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.476060 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.477196 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.477790 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.479400 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqcgf\" (UniqueName: \"kubernetes.io/projected/cce21119-33b5-4f7c-834c-0e8f16ccd825-kube-api-access-mqcgf\") pod \"glance-operator-controller-manager-5568b5d68-25cxf\" (UID: \"cce21119-33b5-4f7c-834c-0e8f16ccd825\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.480823 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-pd586" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.497243 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.498240 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.498290 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.504183 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-zdv44" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.518671 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.527008 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.527693 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.534469 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.535534 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.541254 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-d8vlf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549286 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pgv4\" (UniqueName: \"kubernetes.io/projected/05b084da-1f15-4dba-8d4e-b540d4447318-kube-api-access-2pgv4\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c76wv\" (UniqueName: \"kubernetes.io/projected/45344ba0-bd44-4a54-9628-460586eae8df-kube-api-access-c76wv\") pod \"keystone-operator-controller-manager-655d88ccb9-79ksl\" (UID: \"45344ba0-bd44-4a54-9628-460586eae8df\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549363 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nssvh\" (UniqueName: \"kubernetes.io/projected/70bea185-7eb6-4fc4-8016-c32e11078df1-kube-api-access-nssvh\") pod \"nova-operator-controller-manager-7c7fc454ff-997ct\" (UID: \"70bea185-7eb6-4fc4-8016-c32e11078df1\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr8q4\" (UniqueName: \"kubernetes.io/projected/b89d614b-bb9f-4ba5-bb16-d9fe9731279f-kube-api-access-rr8q4\") pod \"ironic-operator-controller-manager-699b87f775-mrz27\" (UID: \"b89d614b-bb9f-4ba5-bb16-d9fe9731279f\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549467 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwrst\" (UniqueName: \"kubernetes.io/projected/290501cd-b2bb-4520-b8b0-0197017e61c6-kube-api-access-gwrst\") pod \"neutron-operator-controller-manager-8d984cc4d-fff4f\" (UID: \"290501cd-b2bb-4520-b8b0-0197017e61c6\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549492 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htzxl\" (UniqueName: \"kubernetes.io/projected/9acc4601-f863-47b9-8a11-d91615ab1af5-kube-api-access-htzxl\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-m78db\" (UID: \"9acc4601-f863-47b9-8a11-d91615ab1af5\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549510 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8kgl\" (UniqueName: \"kubernetes.io/projected/9477ed03-7f88-4abf-bbcd-65a2888aff71-kube-api-access-q8kgl\") pod \"manila-operator-controller-manager-65d89cfd9f-gmfps\" (UID: \"9477ed03-7f88-4abf-bbcd-65a2888aff71\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549537 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.549552 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k2t6\" (UniqueName: \"kubernetes.io/projected/b1667c4d-4d3a-41b1-a02a-890a2b0f26ea-kube-api-access-7k2t6\") pod \"horizon-operator-controller-manager-54876c876f-dqjpg\" (UID: \"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:38 crc kubenswrapper[4989]: E1006 08:58:38.549983 4989 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 06 08:58:38 crc kubenswrapper[4989]: E1006 08:58:38.550022 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert podName:05b084da-1f15-4dba-8d4e-b540d4447318 nodeName:}" failed. No retries permitted until 2025-10-06 08:58:39.050008162 +0000 UTC m=+1169.840033742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert") pod "infra-operator-controller-manager-658588b8c9-cdzbs" (UID: "05b084da-1f15-4dba-8d4e-b540d4447318") : secret "infra-operator-webhook-server-cert" not found Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.550183 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.553432 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.554393 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.567904 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.570670 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.572014 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.574142 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.574273 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9cmjh" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.582270 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-rc5r5" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.583679 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c76wv\" (UniqueName: \"kubernetes.io/projected/45344ba0-bd44-4a54-9628-460586eae8df-kube-api-access-c76wv\") pod \"keystone-operator-controller-manager-655d88ccb9-79ksl\" (UID: \"45344ba0-bd44-4a54-9628-460586eae8df\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.584497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pgv4\" (UniqueName: \"kubernetes.io/projected/05b084da-1f15-4dba-8d4e-b540d4447318-kube-api-access-2pgv4\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.585178 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k2t6\" (UniqueName: \"kubernetes.io/projected/b1667c4d-4d3a-41b1-a02a-890a2b0f26ea-kube-api-access-7k2t6\") pod \"horizon-operator-controller-manager-54876c876f-dqjpg\" (UID: \"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.589434 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.593243 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr8q4\" (UniqueName: \"kubernetes.io/projected/b89d614b-bb9f-4ba5-bb16-d9fe9731279f-kube-api-access-rr8q4\") pod \"ironic-operator-controller-manager-699b87f775-mrz27\" (UID: \"b89d614b-bb9f-4ba5-bb16-d9fe9731279f\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.593773 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.596971 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-fgtxz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.611627 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.611686 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.613166 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.614559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.616277 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.616561 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lsjnb" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.663075 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666414 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmkrd\" (UniqueName: \"kubernetes.io/projected/97f22726-47c7-47db-b9a2-9efbddfaa7ba-kube-api-access-qmkrd\") pod \"placement-operator-controller-manager-54689d9f88-l62dg\" (UID: \"97f22726-47c7-47db-b9a2-9efbddfaa7ba\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666576 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnqw6\" (UniqueName: \"kubernetes.io/projected/522e1bb1-d22f-4256-b0f0-ae27447bd00a-kube-api-access-dnqw6\") pod \"swift-operator-controller-manager-6859f9b676-kgbzk\" (UID: \"522e1bb1-d22f-4256-b0f0-ae27447bd00a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhpnf\" (UniqueName: \"kubernetes.io/projected/3af9ba41-3325-4e75-a392-306b9999606b-kube-api-access-lhpnf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666763 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwrst\" (UniqueName: \"kubernetes.io/projected/290501cd-b2bb-4520-b8b0-0197017e61c6-kube-api-access-gwrst\") pod \"neutron-operator-controller-manager-8d984cc4d-fff4f\" (UID: \"290501cd-b2bb-4520-b8b0-0197017e61c6\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htzxl\" (UniqueName: \"kubernetes.io/projected/9acc4601-f863-47b9-8a11-d91615ab1af5-kube-api-access-htzxl\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-m78db\" (UID: \"9acc4601-f863-47b9-8a11-d91615ab1af5\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.666957 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8kgl\" (UniqueName: \"kubernetes.io/projected/9477ed03-7f88-4abf-bbcd-65a2888aff71-kube-api-access-q8kgl\") pod \"manila-operator-controller-manager-65d89cfd9f-gmfps\" (UID: \"9477ed03-7f88-4abf-bbcd-65a2888aff71\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.667133 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvq95\" (UniqueName: \"kubernetes.io/projected/9d82b2b6-9374-4cbc-a72c-50aa000a0d52-kube-api-access-hvq95\") pod \"octavia-operator-controller-manager-7468f855d8-jgjms\" (UID: \"9d82b2b6-9374-4cbc-a72c-50aa000a0d52\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.667248 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.667303 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6kc6\" (UniqueName: \"kubernetes.io/projected/21ae0953-021f-40d1-aab0-2b693f29212a-kube-api-access-l6kc6\") pod \"ovn-operator-controller-manager-579449c7d5-lzjq2\" (UID: \"21ae0953-021f-40d1-aab0-2b693f29212a\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.667580 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nssvh\" (UniqueName: \"kubernetes.io/projected/70bea185-7eb6-4fc4-8016-c32e11078df1-kube-api-access-nssvh\") pod \"nova-operator-controller-manager-7c7fc454ff-997ct\" (UID: \"70bea185-7eb6-4fc4-8016-c32e11078df1\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.667710 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.677752 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.706740 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nssvh\" (UniqueName: \"kubernetes.io/projected/70bea185-7eb6-4fc4-8016-c32e11078df1-kube-api-access-nssvh\") pod \"nova-operator-controller-manager-7c7fc454ff-997ct\" (UID: \"70bea185-7eb6-4fc4-8016-c32e11078df1\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.717888 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.724117 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.726138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htzxl\" (UniqueName: \"kubernetes.io/projected/9acc4601-f863-47b9-8a11-d91615ab1af5-kube-api-access-htzxl\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-m78db\" (UID: \"9acc4601-f863-47b9-8a11-d91615ab1af5\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.734395 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.736088 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.763684 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwrst\" (UniqueName: \"kubernetes.io/projected/290501cd-b2bb-4520-b8b0-0197017e61c6-kube-api-access-gwrst\") pod \"neutron-operator-controller-manager-8d984cc4d-fff4f\" (UID: \"290501cd-b2bb-4520-b8b0-0197017e61c6\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.763826 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8kgl\" (UniqueName: \"kubernetes.io/projected/9477ed03-7f88-4abf-bbcd-65a2888aff71-kube-api-access-q8kgl\") pod \"manila-operator-controller-manager-65d89cfd9f-gmfps\" (UID: \"9477ed03-7f88-4abf-bbcd-65a2888aff71\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.766806 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5ptvk" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.771404 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmkrd\" (UniqueName: \"kubernetes.io/projected/97f22726-47c7-47db-b9a2-9efbddfaa7ba-kube-api-access-qmkrd\") pod \"placement-operator-controller-manager-54689d9f88-l62dg\" (UID: \"97f22726-47c7-47db-b9a2-9efbddfaa7ba\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.771678 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnqw6\" (UniqueName: \"kubernetes.io/projected/522e1bb1-d22f-4256-b0f0-ae27447bd00a-kube-api-access-dnqw6\") pod \"swift-operator-controller-manager-6859f9b676-kgbzk\" (UID: \"522e1bb1-d22f-4256-b0f0-ae27447bd00a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.771782 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhpnf\" (UniqueName: \"kubernetes.io/projected/3af9ba41-3325-4e75-a392-306b9999606b-kube-api-access-lhpnf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.771930 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvq95\" (UniqueName: \"kubernetes.io/projected/9d82b2b6-9374-4cbc-a72c-50aa000a0d52-kube-api-access-hvq95\") pod \"octavia-operator-controller-manager-7468f855d8-jgjms\" (UID: \"9d82b2b6-9374-4cbc-a72c-50aa000a0d52\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.772018 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.772110 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6kc6\" (UniqueName: \"kubernetes.io/projected/21ae0953-021f-40d1-aab0-2b693f29212a-kube-api-access-l6kc6\") pod \"ovn-operator-controller-manager-579449c7d5-lzjq2\" (UID: \"21ae0953-021f-40d1-aab0-2b693f29212a\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:38 crc kubenswrapper[4989]: E1006 08:58:38.773221 4989 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 08:58:38 crc kubenswrapper[4989]: E1006 08:58:38.773377 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert podName:3af9ba41-3325-4e75-a392-306b9999606b nodeName:}" failed. No retries permitted until 2025-10-06 08:58:39.273345094 +0000 UTC m=+1170.063370674 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" (UID: "3af9ba41-3325-4e75-a392-306b9999606b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.855999 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.878571 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6kc6\" (UniqueName: \"kubernetes.io/projected/21ae0953-021f-40d1-aab0-2b693f29212a-kube-api-access-l6kc6\") pod \"ovn-operator-controller-manager-579449c7d5-lzjq2\" (UID: \"21ae0953-021f-40d1-aab0-2b693f29212a\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.878900 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.879869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mdxd\" (UniqueName: \"kubernetes.io/projected/a3213843-170f-4e4b-b35b-87f424dd2abb-kube-api-access-7mdxd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-nm6z6\" (UID: \"a3213843-170f-4e4b-b35b-87f424dd2abb\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.885126 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmkrd\" (UniqueName: \"kubernetes.io/projected/97f22726-47c7-47db-b9a2-9efbddfaa7ba-kube-api-access-qmkrd\") pod \"placement-operator-controller-manager-54689d9f88-l62dg\" (UID: \"97f22726-47c7-47db-b9a2-9efbddfaa7ba\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.887496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnqw6\" (UniqueName: \"kubernetes.io/projected/522e1bb1-d22f-4256-b0f0-ae27447bd00a-kube-api-access-dnqw6\") pod \"swift-operator-controller-manager-6859f9b676-kgbzk\" (UID: \"522e1bb1-d22f-4256-b0f0-ae27447bd00a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.894383 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.912023 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.913451 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.916408 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.920358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvq95\" (UniqueName: \"kubernetes.io/projected/9d82b2b6-9374-4cbc-a72c-50aa000a0d52-kube-api-access-hvq95\") pod \"octavia-operator-controller-manager-7468f855d8-jgjms\" (UID: \"9d82b2b6-9374-4cbc-a72c-50aa000a0d52\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.923432 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhpnf\" (UniqueName: \"kubernetes.io/projected/3af9ba41-3325-4e75-a392-306b9999606b-kube-api-access-lhpnf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.947116 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7clqz" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.947784 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.985464 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7"] Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.985643 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.986337 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpsl4\" (UniqueName: \"kubernetes.io/projected/d8ec36ee-cada-48de-b02d-69c1c21ae7b2-kube-api-access-tpsl4\") pod \"test-operator-controller-manager-5cd5cb47d7-b5kt7\" (UID: \"d8ec36ee-cada-48de-b02d-69c1c21ae7b2\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:38 crc kubenswrapper[4989]: I1006 08:58:38.986423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mdxd\" (UniqueName: \"kubernetes.io/projected/a3213843-170f-4e4b-b35b-87f424dd2abb-kube-api-access-7mdxd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-nm6z6\" (UID: \"a3213843-170f-4e4b-b35b-87f424dd2abb\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.042116 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.043641 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.048336 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mdxd\" (UniqueName: \"kubernetes.io/projected/a3213843-170f-4e4b-b35b-87f424dd2abb-kube-api-access-7mdxd\") pod \"telemetry-operator-controller-manager-5d4d74dd89-nm6z6\" (UID: \"a3213843-170f-4e4b-b35b-87f424dd2abb\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.048463 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.054416 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2pnvp" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.074480 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.087618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpsl4\" (UniqueName: \"kubernetes.io/projected/d8ec36ee-cada-48de-b02d-69c1c21ae7b2-kube-api-access-tpsl4\") pod \"test-operator-controller-manager-5cd5cb47d7-b5kt7\" (UID: \"d8ec36ee-cada-48de-b02d-69c1c21ae7b2\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.091403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.091539 4989 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.091574 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert podName:05b084da-1f15-4dba-8d4e-b540d4447318 nodeName:}" failed. No retries permitted until 2025-10-06 08:58:40.091561673 +0000 UTC m=+1170.881587253 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert") pod "infra-operator-controller-manager-658588b8c9-cdzbs" (UID: "05b084da-1f15-4dba-8d4e-b540d4447318") : secret "infra-operator-webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.120157 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.121557 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.126312 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpsl4\" (UniqueName: \"kubernetes.io/projected/d8ec36ee-cada-48de-b02d-69c1c21ae7b2-kube-api-access-tpsl4\") pod \"test-operator-controller-manager-5cd5cb47d7-b5kt7\" (UID: \"d8ec36ee-cada-48de-b02d-69c1c21ae7b2\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.127158 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.127339 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-v44ws" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.132035 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.155807 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.157060 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.159823 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.165622 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8zbhc" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.175313 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.194057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59s9w\" (UniqueName: \"kubernetes.io/projected/428077cb-5a71-4f65-a37e-2384c4c0af19-kube-api-access-59s9w\") pod \"watcher-operator-controller-manager-6cbc6dd547-672zz\" (UID: \"428077cb-5a71-4f65-a37e-2384c4c0af19\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.194135 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.194160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrfq\" (UniqueName: \"kubernetes.io/projected/7fba2f19-add5-4c43-a142-82dd5c0ba564-kube-api-access-plrfq\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.203720 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.219416 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.283677 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.297615 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8qvn\" (UniqueName: \"kubernetes.io/projected/9168e44a-597f-4623-9520-b9fe4c46db99-kube-api-access-z8qvn\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kzsql\" (UID: \"9168e44a-597f-4623-9520-b9fe4c46db99\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.297676 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrfq\" (UniqueName: \"kubernetes.io/projected/7fba2f19-add5-4c43-a142-82dd5c0ba564-kube-api-access-plrfq\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.297797 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.297851 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59s9w\" (UniqueName: \"kubernetes.io/projected/428077cb-5a71-4f65-a37e-2384c4c0af19-kube-api-access-59s9w\") pod \"watcher-operator-controller-manager-6cbc6dd547-672zz\" (UID: \"428077cb-5a71-4f65-a37e-2384c4c0af19\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.297873 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.297984 4989 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.298030 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert podName:7fba2f19-add5-4c43-a142-82dd5c0ba564 nodeName:}" failed. No retries permitted until 2025-10-06 08:58:39.798015247 +0000 UTC m=+1170.588040827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert") pod "openstack-operator-controller-manager-669d7f654d-rvb7c" (UID: "7fba2f19-add5-4c43-a142-82dd5c0ba564") : secret "webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.299057 4989 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: E1006 08:58:39.299114 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert podName:3af9ba41-3325-4e75-a392-306b9999606b nodeName:}" failed. No retries permitted until 2025-10-06 08:58:40.299097829 +0000 UTC m=+1171.089123409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" (UID: "3af9ba41-3325-4e75-a392-306b9999606b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.319847 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrfq\" (UniqueName: \"kubernetes.io/projected/7fba2f19-add5-4c43-a142-82dd5c0ba564-kube-api-access-plrfq\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.336924 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59s9w\" (UniqueName: \"kubernetes.io/projected/428077cb-5a71-4f65-a37e-2384c4c0af19-kube-api-access-59s9w\") pod \"watcher-operator-controller-manager-6cbc6dd547-672zz\" (UID: \"428077cb-5a71-4f65-a37e-2384c4c0af19\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.337346 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.376638 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:39 crc kubenswrapper[4989]: W1006 08:58:39.393842 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e971444_f7c2_4443_ad67_8344cc2ece01.slice/crio-e4c18d81f1878926bdb31cf7dcb12df10dce946879930d4f737c1a1646df2bda WatchSource:0}: Error finding container e4c18d81f1878926bdb31cf7dcb12df10dce946879930d4f737c1a1646df2bda: Status 404 returned error can't find the container with id e4c18d81f1878926bdb31cf7dcb12df10dce946879930d4f737c1a1646df2bda Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.401715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8qvn\" (UniqueName: \"kubernetes.io/projected/9168e44a-597f-4623-9520-b9fe4c46db99-kube-api-access-z8qvn\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kzsql\" (UID: \"9168e44a-597f-4623-9520-b9fe4c46db99\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.422297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8qvn\" (UniqueName: \"kubernetes.io/projected/9168e44a-597f-4623-9520-b9fe4c46db99-kube-api-access-z8qvn\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-kzsql\" (UID: \"9168e44a-597f-4623-9520-b9fe4c46db99\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.553238 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.638040 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" event={"ID":"9e971444-f7c2-4443-ad67-8344cc2ece01","Type":"ContainerStarted","Data":"e4c18d81f1878926bdb31cf7dcb12df10dce946879930d4f737c1a1646df2bda"} Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.723067 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx"] Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.741611 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5"] Oct 06 08:58:39 crc kubenswrapper[4989]: W1006 08:58:39.773001 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ed264b3_7650_41ed_bb45_c5e0265ebcaf.slice/crio-871b59d2aaaed110cd62152ea1ee5cfabe977ce5f1949a5a6a13d1be53aa3817 WatchSource:0}: Error finding container 871b59d2aaaed110cd62152ea1ee5cfabe977ce5f1949a5a6a13d1be53aa3817: Status 404 returned error can't find the container with id 871b59d2aaaed110cd62152ea1ee5cfabe977ce5f1949a5a6a13d1be53aa3817 Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.808203 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:39 crc kubenswrapper[4989]: I1006 08:58:39.813193 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fba2f19-add5-4c43-a142-82dd5c0ba564-cert\") pod \"openstack-operator-controller-manager-669d7f654d-rvb7c\" (UID: \"7fba2f19-add5-4c43-a142-82dd5c0ba564\") " pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.082037 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.087146 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30e9d8ad_e772_44e4_8234_a0ef8a3483e1.slice/crio-93a0e8a4344d163002bf7dd03d142477308b77fd7909521302396c054cc45f23 WatchSource:0}: Error finding container 93a0e8a4344d163002bf7dd03d142477308b77fd7909521302396c054cc45f23: Status 404 returned error can't find the container with id 93a0e8a4344d163002bf7dd03d142477308b77fd7909521302396c054cc45f23 Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.088297 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.091721 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.112901 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.116072 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.122237 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05b084da-1f15-4dba-8d4e-b540d4447318-cert\") pod \"infra-operator-controller-manager-658588b8c9-cdzbs\" (UID: \"05b084da-1f15-4dba-8d4e-b540d4447318\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.128750 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.133072 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1667c4d_4d3a_41b1_a02a_890a2b0f26ea.slice/crio-8456118cf90774b2752dae6b86ee7b4a8dd71b25421204d7ccb64adc95114a7e WatchSource:0}: Error finding container 8456118cf90774b2752dae6b86ee7b4a8dd71b25421204d7ccb64adc95114a7e: Status 404 returned error can't find the container with id 8456118cf90774b2752dae6b86ee7b4a8dd71b25421204d7ccb64adc95114a7e Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.136696 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.140174 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9acc4601_f863_47b9_8a11_d91615ab1af5.slice/crio-26a37ebb6198fdf8db98c25c92df6b0d927c462ccd78f057bb6d195ba415226d WatchSource:0}: Error finding container 26a37ebb6198fdf8db98c25c92df6b0d927c462ccd78f057bb6d195ba415226d: Status 404 returned error can't find the container with id 26a37ebb6198fdf8db98c25c92df6b0d927c462ccd78f057bb6d195ba415226d Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.184680 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.184790 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8ec36ee_cada_48de_b02d_69c1c21ae7b2.slice/crio-abcc649f9b56adc8bce8733e4ed891d4084cdaf69319b5a423a2a51389865337 WatchSource:0}: Error finding container abcc649f9b56adc8bce8733e4ed891d4084cdaf69319b5a423a2a51389865337: Status 404 returned error can't find the container with id abcc649f9b56adc8bce8733e4ed891d4084cdaf69319b5a423a2a51389865337 Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.195194 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.206090 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod290501cd_b2bb_4520_b8b0_0197017e61c6.slice/crio-b7c9526dd1dbffeab5d6ed70d01d7532c2ee86090dbf23958d3b20980c47b6d3 WatchSource:0}: Error finding container b7c9526dd1dbffeab5d6ed70d01d7532c2ee86090dbf23958d3b20980c47b6d3: Status 404 returned error can't find the container with id b7c9526dd1dbffeab5d6ed70d01d7532c2ee86090dbf23958d3b20980c47b6d3 Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.209212 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.216573 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.218800 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb89d614b_bb9f_4ba5_bb16_d9fe9731279f.slice/crio-39c09fbc96c077e46955d10120728084ed65f1617879b42fac8de9cecf05fab9 WatchSource:0}: Error finding container 39c09fbc96c077e46955d10120728084ed65f1617879b42fac8de9cecf05fab9: Status 404 returned error can't find the container with id 39c09fbc96c077e46955d10120728084ed65f1617879b42fac8de9cecf05fab9 Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.272702 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.318346 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.323168 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3af9ba41-3325-4e75-a392-306b9999606b-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz\" (UID: \"3af9ba41-3325-4e75-a392-306b9999606b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.367689 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.527984 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.550110 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.552336 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9168e44a_597f_4623_9520_b9fe4c46db99.slice/crio-73f51ebdca36544e52fd7e5fdff280c98071edc83835a39f75b472dbf59dd13b WatchSource:0}: Error finding container 73f51ebdca36544e52fd7e5fdff280c98071edc83835a39f75b472dbf59dd13b: Status 404 returned error can't find the container with id 73f51ebdca36544e52fd7e5fdff280c98071edc83835a39f75b472dbf59dd13b Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.568037 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.585268 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.585897 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97f22726_47c7_47db_b9a2_9efbddfaa7ba.slice/crio-0c93f1e500eec21b334a0dbc2208bd5d53ff6e026fa28ef46e6ed0219c0f8c37 WatchSource:0}: Error finding container 0c93f1e500eec21b334a0dbc2208bd5d53ff6e026fa28ef46e6ed0219c0f8c37: Status 404 returned error can't find the container with id 0c93f1e500eec21b334a0dbc2208bd5d53ff6e026fa28ef46e6ed0219c0f8c37 Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.590748 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21ae0953_021f_40d1_aab0_2b693f29212a.slice/crio-ddd6fa05eb847aa9296cd3ec5f563c7689738081b8b84c54cc6b37dfde7f4612 WatchSource:0}: Error finding container ddd6fa05eb847aa9296cd3ec5f563c7689738081b8b84c54cc6b37dfde7f4612: Status 404 returned error can't find the container with id ddd6fa05eb847aa9296cd3ec5f563c7689738081b8b84c54cc6b37dfde7f4612 Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.593714 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3213843_170f_4e4b_b35b_87f424dd2abb.slice/crio-f5f81ea643bc2bfcab3bda52fac8ec8974a6ffed9c6895e38dda45554add1cca WatchSource:0}: Error finding container f5f81ea643bc2bfcab3bda52fac8ec8974a6ffed9c6895e38dda45554add1cca: Status 404 returned error can't find the container with id f5f81ea643bc2bfcab3bda52fac8ec8974a6ffed9c6895e38dda45554add1cca Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.598136 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6kc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-lzjq2_openstack-operators(21ae0953-021f-40d1-aab0-2b693f29212a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.604602 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.615462 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.623068 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.627805 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9477ed03_7f88_4abf_bbcd_65a2888aff71.slice/crio-28ea69143ca1d846647010aa41c366ebaf9202cea28e67cfd6cb8c88473dbd33 WatchSource:0}: Error finding container 28ea69143ca1d846647010aa41c366ebaf9202cea28e67cfd6cb8c88473dbd33: Status 404 returned error can't find the container with id 28ea69143ca1d846647010aa41c366ebaf9202cea28e67cfd6cb8c88473dbd33 Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.629250 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-59s9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-672zz_openstack-operators(428077cb-5a71-4f65-a37e-2384c4c0af19): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.629456 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk"] Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.636005 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q8kgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-65d89cfd9f-gmfps_openstack-operators(9477ed03-7f88-4abf-bbcd-65a2888aff71): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.636356 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c"] Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.648272 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" event={"ID":"70bea185-7eb6-4fc4-8016-c32e11078df1","Type":"ContainerStarted","Data":"f26c4127397e800ecc27b2fff24b4d177e876553911a7e9072270d5ba88393c3"} Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.652527 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod522e1bb1_d22f_4256_b0f0_ae27447bd00a.slice/crio-851f28c3748a4fa1c38cc69906c53caf44f12dc85fff426a8639340ab872d78f WatchSource:0}: Error finding container 851f28c3748a4fa1c38cc69906c53caf44f12dc85fff426a8639340ab872d78f: Status 404 returned error can't find the container with id 851f28c3748a4fa1c38cc69906c53caf44f12dc85fff426a8639340ab872d78f Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.653372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" event={"ID":"21ae0953-021f-40d1-aab0-2b693f29212a","Type":"ContainerStarted","Data":"ddd6fa05eb847aa9296cd3ec5f563c7689738081b8b84c54cc6b37dfde7f4612"} Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.655490 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d82b2b6_9374_4cbc_a72c_50aa000a0d52.slice/crio-9b6a415a822ac00332735278c83626fcdb3902c2b148c0cf7ad44983edcb4fed WatchSource:0}: Error finding container 9b6a415a822ac00332735278c83626fcdb3902c2b148c0cf7ad44983edcb4fed: Status 404 returned error can't find the container with id 9b6a415a822ac00332735278c83626fcdb3902c2b148c0cf7ad44983edcb4fed Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.657809 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" event={"ID":"c0f513bb-5736-42bd-a6cb-cf5314b940ee","Type":"ContainerStarted","Data":"364afe698c2b2060af9c30b8cfe877402d7213a323012712b417c7db5274ba78"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.660372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" event={"ID":"d8ec36ee-cada-48de-b02d-69c1c21ae7b2","Type":"ContainerStarted","Data":"abcc649f9b56adc8bce8733e4ed891d4084cdaf69319b5a423a2a51389865337"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.662380 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" event={"ID":"428077cb-5a71-4f65-a37e-2384c4c0af19","Type":"ContainerStarted","Data":"ad775775b9c8b1b0f9965ccab6ecc796e8a98fd2e66979a98c0ebba8d36354cf"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.672174 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" event={"ID":"a3213843-170f-4e4b-b35b-87f424dd2abb","Type":"ContainerStarted","Data":"f5f81ea643bc2bfcab3bda52fac8ec8974a6ffed9c6895e38dda45554add1cca"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.675577 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" event={"ID":"9168e44a-597f-4623-9520-b9fe4c46db99","Type":"ContainerStarted","Data":"73f51ebdca36544e52fd7e5fdff280c98071edc83835a39f75b472dbf59dd13b"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.677844 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" event={"ID":"cce21119-33b5-4f7c-834c-0e8f16ccd825","Type":"ContainerStarted","Data":"8fe10c59d9e305f456f6a0742f1ee164344315a47661d2efe4e4e5ef55dfae06"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.679710 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" event={"ID":"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea","Type":"ContainerStarted","Data":"8456118cf90774b2752dae6b86ee7b4a8dd71b25421204d7ccb64adc95114a7e"} Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.680203 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hvq95,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-jgjms_openstack-operators(9d82b2b6-9374-4cbc-a72c-50aa000a0d52): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.680371 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dnqw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-kgbzk_openstack-operators(522e1bb1-d22f-4256-b0f0-ae27447bd00a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.683578 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" event={"ID":"9477ed03-7f88-4abf-bbcd-65a2888aff71","Type":"ContainerStarted","Data":"28ea69143ca1d846647010aa41c366ebaf9202cea28e67cfd6cb8c88473dbd33"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.687316 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" event={"ID":"45344ba0-bd44-4a54-9628-460586eae8df","Type":"ContainerStarted","Data":"c625a076801cd120703c32b4cad24db39b85d869dd244f2fe8df0ddbdd68b660"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.689071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" event={"ID":"b89d614b-bb9f-4ba5-bb16-d9fe9731279f","Type":"ContainerStarted","Data":"39c09fbc96c077e46955d10120728084ed65f1617879b42fac8de9cecf05fab9"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.696746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" event={"ID":"97f22726-47c7-47db-b9a2-9efbddfaa7ba","Type":"ContainerStarted","Data":"0c93f1e500eec21b334a0dbc2208bd5d53ff6e026fa28ef46e6ed0219c0f8c37"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.698445 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" event={"ID":"30e9d8ad-e772-44e4-8234-a0ef8a3483e1","Type":"ContainerStarted","Data":"93a0e8a4344d163002bf7dd03d142477308b77fd7909521302396c054cc45f23"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.700578 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" event={"ID":"9ed264b3-7650-41ed-bb45-c5e0265ebcaf","Type":"ContainerStarted","Data":"871b59d2aaaed110cd62152ea1ee5cfabe977ce5f1949a5a6a13d1be53aa3817"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.712750 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" event={"ID":"9acc4601-f863-47b9-8a11-d91615ab1af5","Type":"ContainerStarted","Data":"26a37ebb6198fdf8db98c25c92df6b0d927c462ccd78f057bb6d195ba415226d"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.714635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" event={"ID":"290501cd-b2bb-4520-b8b0-0197017e61c6","Type":"ContainerStarted","Data":"b7c9526dd1dbffeab5d6ed70d01d7532c2ee86090dbf23958d3b20980c47b6d3"} Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.758467 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.787784 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05b084da_1f15_4dba_8d4e_b540d4447318.slice/crio-1759ddc113dd580174641e7ecafbae34a9a883190f732b5c5d4a35cb30400a1d WatchSource:0}: Error finding container 1759ddc113dd580174641e7ecafbae34a9a883190f732b5c5d4a35cb30400a1d: Status 404 returned error can't find the container with id 1759ddc113dd580174641e7ecafbae34a9a883190f732b5c5d4a35cb30400a1d Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.813543 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" podUID="21ae0953-021f-40d1-aab0-2b693f29212a" Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.876598 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" podUID="428077cb-5a71-4f65-a37e-2384c4c0af19" Oct 06 08:58:40 crc kubenswrapper[4989]: I1006 08:58:40.927411 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz"] Oct 06 08:58:40 crc kubenswrapper[4989]: W1006 08:58:40.947574 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3af9ba41_3325_4e75_a392_306b9999606b.slice/crio-f2154ec2ab48bc3ff57a5abb7e23c29ebcac3841cfd43c7fc508666a325f2c68 WatchSource:0}: Error finding container f2154ec2ab48bc3ff57a5abb7e23c29ebcac3841cfd43c7fc508666a325f2c68: Status 404 returned error can't find the container with id f2154ec2ab48bc3ff57a5abb7e23c29ebcac3841cfd43c7fc508666a325f2c68 Oct 06 08:58:40 crc kubenswrapper[4989]: E1006 08:58:40.950088 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhpnf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz_openstack-operators(3af9ba41-3325-4e75-a392-306b9999606b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.042009 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" podUID="522e1bb1-d22f-4256-b0f0-ae27447bd00a" Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.044536 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" podUID="9477ed03-7f88-4abf-bbcd-65a2888aff71" Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.044637 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" podUID="9d82b2b6-9374-4cbc-a72c-50aa000a0d52" Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.140163 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" podUID="3af9ba41-3325-4e75-a392-306b9999606b" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.748511 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" event={"ID":"9477ed03-7f88-4abf-bbcd-65a2888aff71","Type":"ContainerStarted","Data":"61ac44f2c68cda71c1abacbe0d7e4e220c63fbe886afe4bd751c933003c4a214"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.751157 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" event={"ID":"05b084da-1f15-4dba-8d4e-b540d4447318","Type":"ContainerStarted","Data":"1759ddc113dd580174641e7ecafbae34a9a883190f732b5c5d4a35cb30400a1d"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.752217 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757\\\"\"" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" podUID="9477ed03-7f88-4abf-bbcd-65a2888aff71" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.757092 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" event={"ID":"7fba2f19-add5-4c43-a142-82dd5c0ba564","Type":"ContainerStarted","Data":"b4e801ba341a0dff4c2700ecec9063656ae7a77ec35e5075a66702f9546532cf"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.757125 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" event={"ID":"7fba2f19-add5-4c43-a142-82dd5c0ba564","Type":"ContainerStarted","Data":"2fa7c844fa807e43dfa1578a1019bd9a5e8ff87bb0e19061be54af3fe8c93161"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.757137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" event={"ID":"7fba2f19-add5-4c43-a142-82dd5c0ba564","Type":"ContainerStarted","Data":"e2bad1d72b1912f38a2712284d2e268e1f8f32db6ad44188ba703b88b2746516"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.757741 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.761362 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" event={"ID":"9d82b2b6-9374-4cbc-a72c-50aa000a0d52","Type":"ContainerStarted","Data":"5a00b152449703f97d248895e38e66f23168497b97786394cdcc3f12ff345146"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.761441 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" event={"ID":"9d82b2b6-9374-4cbc-a72c-50aa000a0d52","Type":"ContainerStarted","Data":"9b6a415a822ac00332735278c83626fcdb3902c2b148c0cf7ad44983edcb4fed"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.763868 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" podUID="9d82b2b6-9374-4cbc-a72c-50aa000a0d52" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.767359 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" event={"ID":"522e1bb1-d22f-4256-b0f0-ae27447bd00a","Type":"ContainerStarted","Data":"ebfd43883f48eb39737993639a11653bd5a6a49611d427875ba18432e91e3119"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.767446 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" event={"ID":"522e1bb1-d22f-4256-b0f0-ae27447bd00a","Type":"ContainerStarted","Data":"851f28c3748a4fa1c38cc69906c53caf44f12dc85fff426a8639340ab872d78f"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.770176 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" podUID="522e1bb1-d22f-4256-b0f0-ae27447bd00a" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.773380 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" event={"ID":"21ae0953-021f-40d1-aab0-2b693f29212a","Type":"ContainerStarted","Data":"f035e465547f184b3a1221ab7a55f20a01375f14fe8423ac5d7dde2f6cf98d81"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.774841 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" podUID="21ae0953-021f-40d1-aab0-2b693f29212a" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.776615 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" event={"ID":"3af9ba41-3325-4e75-a392-306b9999606b","Type":"ContainerStarted","Data":"0e430ec998645c1c8afeb91e118c2d207202529255950781735c325b35b67fc6"} Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.776718 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" event={"ID":"3af9ba41-3325-4e75-a392-306b9999606b","Type":"ContainerStarted","Data":"f2154ec2ab48bc3ff57a5abb7e23c29ebcac3841cfd43c7fc508666a325f2c68"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.778773 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" podUID="3af9ba41-3325-4e75-a392-306b9999606b" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.798394 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" podStartSLOduration=3.798377327 podStartE2EDuration="3.798377327s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:58:41.798353816 +0000 UTC m=+1172.588379396" watchObservedRunningTime="2025-10-06 08:58:41.798377327 +0000 UTC m=+1172.588402907" Oct 06 08:58:41 crc kubenswrapper[4989]: I1006 08:58:41.803472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" event={"ID":"428077cb-5a71-4f65-a37e-2384c4c0af19","Type":"ContainerStarted","Data":"bbc4449d34f88ae30ee1298f273362b401e038dfcb8a8aedccab59244d719cc1"} Oct 06 08:58:41 crc kubenswrapper[4989]: E1006 08:58:41.807056 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" podUID="428077cb-5a71-4f65-a37e-2384c4c0af19" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.816326 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" podUID="428077cb-5a71-4f65-a37e-2384c4c0af19" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.816602 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bcd1acac74e68eea5a9c3b7ba1bcb29d3a5b43423fc23c19ad4715bdac41f799\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" podUID="3af9ba41-3325-4e75-a392-306b9999606b" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.816696 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:063aae1458289d1090a77c74c2b978b9eb978b0e4062c399f0cb5434a8dd2757\\\"\"" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" podUID="9477ed03-7f88-4abf-bbcd-65a2888aff71" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.816727 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" podUID="9d82b2b6-9374-4cbc-a72c-50aa000a0d52" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.816756 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" podUID="522e1bb1-d22f-4256-b0f0-ae27447bd00a" Oct 06 08:58:42 crc kubenswrapper[4989]: E1006 08:58:42.818281 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" podUID="21ae0953-021f-40d1-aab0-2b693f29212a" Oct 06 08:58:50 crc kubenswrapper[4989]: I1006 08:58:50.098729 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-669d7f654d-rvb7c" Oct 06 08:58:51 crc kubenswrapper[4989]: I1006 08:58:51.893158 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" event={"ID":"9e971444-f7c2-4443-ad67-8344cc2ece01","Type":"ContainerStarted","Data":"f2644a54905d85e09704b8cbbbfd1aff6967d0348bdb16f7ff5b3dacc8063e49"} Oct 06 08:58:51 crc kubenswrapper[4989]: I1006 08:58:51.896088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" event={"ID":"97f22726-47c7-47db-b9a2-9efbddfaa7ba","Type":"ContainerStarted","Data":"78bc21cba3286eda43ad2faeae37fad218f9aa50585a8f3cce609e3b41a4cf4a"} Oct 06 08:58:51 crc kubenswrapper[4989]: I1006 08:58:51.898762 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" event={"ID":"c0f513bb-5736-42bd-a6cb-cf5314b940ee","Type":"ContainerStarted","Data":"5e598e06443d12ab05e6873508a294797df3a7ce10c8896d316f265518d15488"} Oct 06 08:58:51 crc kubenswrapper[4989]: I1006 08:58:51.906241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" event={"ID":"9ed264b3-7650-41ed-bb45-c5e0265ebcaf","Type":"ContainerStarted","Data":"8b8c38da3bfbd658a3a74559cc258b9938b262125be690be5b058ebd7c1c0952"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.940645 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" event={"ID":"97f22726-47c7-47db-b9a2-9efbddfaa7ba","Type":"ContainerStarted","Data":"66946830daed76f6595a2b146d2c3a92fb690a8290d3b9ac2041afad85322476"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.940944 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.953875 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" event={"ID":"70bea185-7eb6-4fc4-8016-c32e11078df1","Type":"ContainerStarted","Data":"3d43a25d46b8d325caa06a50e3e913e4773cbe99844e25e42c6f8099b5374643"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.955419 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" event={"ID":"9e971444-f7c2-4443-ad67-8344cc2ece01","Type":"ContainerStarted","Data":"b5bc8f72622e912855e87830434c36022831eb53657f77d5ee920ab7e274dec5"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.955602 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.957106 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" event={"ID":"290501cd-b2bb-4520-b8b0-0197017e61c6","Type":"ContainerStarted","Data":"8c5f511646a96bca5f16972f88979ed5f73f393fa0cf7fe908c8e7ec287cff53"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.957132 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" event={"ID":"290501cd-b2bb-4520-b8b0-0197017e61c6","Type":"ContainerStarted","Data":"599ab06e1964f1ddff43427901c19245757c28a3324d634ead04962b5e48593b"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.957859 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.960187 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" event={"ID":"9168e44a-597f-4623-9520-b9fe4c46db99","Type":"ContainerStarted","Data":"7219fa5779c761eef3d7228db731b6fd47c95fbcf6737accc1ab872444d1a34b"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.964902 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" event={"ID":"cce21119-33b5-4f7c-834c-0e8f16ccd825","Type":"ContainerStarted","Data":"03d1110718c47f54fe1b1f35f404979a63bf88d32011dcbfafd6ca909451827e"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.964944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" event={"ID":"cce21119-33b5-4f7c-834c-0e8f16ccd825","Type":"ContainerStarted","Data":"bde366775927bd19b88602ecd4a6526cfa24454f9fc0ddd9accd3876289c0de0"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.965633 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.970136 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" podStartSLOduration=4.177450141 podStartE2EDuration="14.97012255s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.590166898 +0000 UTC m=+1171.380192478" lastFinishedPulling="2025-10-06 08:58:51.382839297 +0000 UTC m=+1182.172864887" observedRunningTime="2025-10-06 08:58:52.966891257 +0000 UTC m=+1183.756916837" watchObservedRunningTime="2025-10-06 08:58:52.97012255 +0000 UTC m=+1183.760148130" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.974082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" event={"ID":"d8ec36ee-cada-48de-b02d-69c1c21ae7b2","Type":"ContainerStarted","Data":"9e018451b2903823a91bdbe6191fdee9e26433891a494a585c88612a78076b5b"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.978241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" event={"ID":"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea","Type":"ContainerStarted","Data":"42040e4ade8b6e81c3dde8ed7b7d182dfa97c864c8110efba629ba3500aeeeef"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.978296 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" event={"ID":"b1667c4d-4d3a-41b1-a02a-890a2b0f26ea","Type":"ContainerStarted","Data":"91b4f839bb705739f5001304d82560c51cc29e13fda0e7f2ef5670910a184324"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.978984 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.982470 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" event={"ID":"9ed264b3-7650-41ed-bb45-c5e0265ebcaf","Type":"ContainerStarted","Data":"baacdc6cf32a180a543fda3cf4626db0a25a7705f5d6767c0a088ad16519027e"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.983219 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.986379 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" event={"ID":"c0f513bb-5736-42bd-a6cb-cf5314b940ee","Type":"ContainerStarted","Data":"f7da4994901f98d291b7b2f0747b5620600e17b0ac9b20a6339cfaf6ca3f0e8d"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.987143 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.992391 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" event={"ID":"b89d614b-bb9f-4ba5-bb16-d9fe9731279f","Type":"ContainerStarted","Data":"8f113568e5ab4c00ff33ee17a4a7b986cf5eedec138ccf9e475715a005f94a4e"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.993405 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" podStartSLOduration=3.728845671 podStartE2EDuration="14.993384281s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.129572493 +0000 UTC m=+1170.919598073" lastFinishedPulling="2025-10-06 08:58:51.394111093 +0000 UTC m=+1182.184136683" observedRunningTime="2025-10-06 08:58:52.990677913 +0000 UTC m=+1183.780703493" watchObservedRunningTime="2025-10-06 08:58:52.993384281 +0000 UTC m=+1183.783409861" Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.994624 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" event={"ID":"05b084da-1f15-4dba-8d4e-b540d4447318","Type":"ContainerStarted","Data":"1ef00f646a015ea13c1831ce71f4a1326010ba3608334759a31af84d6eb69003"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.994671 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" event={"ID":"05b084da-1f15-4dba-8d4e-b540d4447318","Type":"ContainerStarted","Data":"9d6cca3a8db98f8645f77881b94d15b73a756513bf076255a2d153ded893f262"} Oct 06 08:58:52 crc kubenswrapper[4989]: I1006 08:58:52.995039 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.008276 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" event={"ID":"45344ba0-bd44-4a54-9628-460586eae8df","Type":"ContainerStarted","Data":"d671d01be24299ce879c1aca8211cf0679277484f0a787517dfb3fe44469edfb"} Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.022926 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-kzsql" podStartSLOduration=3.14413186 podStartE2EDuration="14.022903442s" podCreationTimestamp="2025-10-06 08:58:39 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.567097953 +0000 UTC m=+1171.357123533" lastFinishedPulling="2025-10-06 08:58:51.445869535 +0000 UTC m=+1182.235895115" observedRunningTime="2025-10-06 08:58:53.012907554 +0000 UTC m=+1183.802933134" watchObservedRunningTime="2025-10-06 08:58:53.022903442 +0000 UTC m=+1183.812929012" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.026412 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" event={"ID":"30e9d8ad-e772-44e4-8234-a0ef8a3483e1","Type":"ContainerStarted","Data":"4db4c932ef90de4ec2a3bb340e2958926c8cce320cada2e3494656987d7143a8"} Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.026458 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" event={"ID":"30e9d8ad-e772-44e4-8234-a0ef8a3483e1","Type":"ContainerStarted","Data":"c0b49bfdd2938ecddadd05907469998c040f399fd8473a3f0ba55c40c1353175"} Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.027087 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.040081 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" event={"ID":"a3213843-170f-4e4b-b35b-87f424dd2abb","Type":"ContainerStarted","Data":"f811d51c3fc34794bf0be28333bd99b383b9e7d281d18175c1d2d2f71e74e98a"} Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.040699 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.052061 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" podStartSLOduration=3.075885647 podStartE2EDuration="15.052045203s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:39.397715893 +0000 UTC m=+1170.187741473" lastFinishedPulling="2025-10-06 08:58:51.373875429 +0000 UTC m=+1182.163901029" observedRunningTime="2025-10-06 08:58:53.051302101 +0000 UTC m=+1183.841327671" watchObservedRunningTime="2025-10-06 08:58:53.052045203 +0000 UTC m=+1183.842070783" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.059032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" event={"ID":"9acc4601-f863-47b9-8a11-d91615ab1af5","Type":"ContainerStarted","Data":"3749a958f82b4783dcfabb0e57fdac2c2bcd8b3c140a533cb2e58bd555af0f4d"} Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.082402 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" podStartSLOduration=3.8483575979999998 podStartE2EDuration="15.082384588s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.20818159 +0000 UTC m=+1170.998207170" lastFinishedPulling="2025-10-06 08:58:51.44220858 +0000 UTC m=+1182.232234160" observedRunningTime="2025-10-06 08:58:53.074987855 +0000 UTC m=+1183.865013435" watchObservedRunningTime="2025-10-06 08:58:53.082384588 +0000 UTC m=+1183.872410168" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.107767 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" podStartSLOduration=4.485273358 podStartE2EDuration="15.107741369s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.799202297 +0000 UTC m=+1171.589227877" lastFinishedPulling="2025-10-06 08:58:51.421670308 +0000 UTC m=+1182.211695888" observedRunningTime="2025-10-06 08:58:53.099606415 +0000 UTC m=+1183.889631995" watchObservedRunningTime="2025-10-06 08:58:53.107741369 +0000 UTC m=+1183.897766949" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.123951 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" podStartSLOduration=3.4902390690000002 podStartE2EDuration="15.123931126s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:39.759231541 +0000 UTC m=+1170.549257121" lastFinishedPulling="2025-10-06 08:58:51.392923598 +0000 UTC m=+1182.182949178" observedRunningTime="2025-10-06 08:58:53.118472319 +0000 UTC m=+1183.908497899" watchObservedRunningTime="2025-10-06 08:58:53.123931126 +0000 UTC m=+1183.913956706" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.159927 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" podStartSLOduration=4.310410986 podStartE2EDuration="15.159910554s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.596368277 +0000 UTC m=+1171.386393847" lastFinishedPulling="2025-10-06 08:58:51.445867835 +0000 UTC m=+1182.235893415" observedRunningTime="2025-10-06 08:58:53.153633923 +0000 UTC m=+1183.943659503" watchObservedRunningTime="2025-10-06 08:58:53.159910554 +0000 UTC m=+1183.949936134" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.180630 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" podStartSLOduration=3.543384391 podStartE2EDuration="15.180615171s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:39.780531225 +0000 UTC m=+1170.570556805" lastFinishedPulling="2025-10-06 08:58:51.417762005 +0000 UTC m=+1182.207787585" observedRunningTime="2025-10-06 08:58:53.176019839 +0000 UTC m=+1183.966045409" watchObservedRunningTime="2025-10-06 08:58:53.180615171 +0000 UTC m=+1183.970640741" Oct 06 08:58:53 crc kubenswrapper[4989]: I1006 08:58:53.203452 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" podStartSLOduration=3.959500274 podStartE2EDuration="15.20343833s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.138054277 +0000 UTC m=+1170.928079857" lastFinishedPulling="2025-10-06 08:58:51.381992323 +0000 UTC m=+1182.172017913" observedRunningTime="2025-10-06 08:58:53.200542196 +0000 UTC m=+1183.990567776" watchObservedRunningTime="2025-10-06 08:58:53.20343833 +0000 UTC m=+1183.993463910" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.068385 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" event={"ID":"b89d614b-bb9f-4ba5-bb16-d9fe9731279f","Type":"ContainerStarted","Data":"6977d2749913ea30edf253602fb5e1626a208dedb7c1371dd589cef9c48dff88"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.068467 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.071048 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" event={"ID":"70bea185-7eb6-4fc4-8016-c32e11078df1","Type":"ContainerStarted","Data":"a9c638fe309ba00ba0b6e52d09240ee08a10ad26cc0ede070cb1398d1912f123"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.071304 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.073045 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" event={"ID":"d8ec36ee-cada-48de-b02d-69c1c21ae7b2","Type":"ContainerStarted","Data":"3d566b44f4a47108661b349b97179e17ca9348b341e2e3b35b7bf2fbecdbc013"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.073192 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.074991 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" event={"ID":"a3213843-170f-4e4b-b35b-87f424dd2abb","Type":"ContainerStarted","Data":"a4992eca7a34c61b0156600dee348ec00227140b7764e231a6c3133359590553"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.077158 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" event={"ID":"9acc4601-f863-47b9-8a11-d91615ab1af5","Type":"ContainerStarted","Data":"fedc6e3dcc71432a040de474be0530bc82e20d55743442f05f22dbf1594200ac"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.077288 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.080086 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" event={"ID":"45344ba0-bd44-4a54-9628-460586eae8df","Type":"ContainerStarted","Data":"aa3ee6d6719bd0fc7ed4358de5aad915115ef4cf49dbc71e5a1ff1992c3ceb6a"} Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.081941 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.092493 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" podStartSLOduration=4.894072621 podStartE2EDuration="16.092470893s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.220231438 +0000 UTC m=+1171.010257018" lastFinishedPulling="2025-10-06 08:58:51.41862971 +0000 UTC m=+1182.208655290" observedRunningTime="2025-10-06 08:58:54.088505408 +0000 UTC m=+1184.878531008" watchObservedRunningTime="2025-10-06 08:58:54.092470893 +0000 UTC m=+1184.882496473" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.097735 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" podStartSLOduration=4.821146797 podStartE2EDuration="16.097716974s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.105411046 +0000 UTC m=+1170.895436626" lastFinishedPulling="2025-10-06 08:58:51.381981223 +0000 UTC m=+1182.172006803" observedRunningTime="2025-10-06 08:58:53.229582944 +0000 UTC m=+1184.019608524" watchObservedRunningTime="2025-10-06 08:58:54.097716974 +0000 UTC m=+1184.887742554" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.118352 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" podStartSLOduration=4.839660962 podStartE2EDuration="16.118330509s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.142673391 +0000 UTC m=+1170.932698971" lastFinishedPulling="2025-10-06 08:58:51.421342938 +0000 UTC m=+1182.211368518" observedRunningTime="2025-10-06 08:58:54.109101092 +0000 UTC m=+1184.899126672" watchObservedRunningTime="2025-10-06 08:58:54.118330509 +0000 UTC m=+1184.908356099" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.126298 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" podStartSLOduration=4.832888856 podStartE2EDuration="16.126278698s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.125210057 +0000 UTC m=+1170.915235647" lastFinishedPulling="2025-10-06 08:58:51.418599909 +0000 UTC m=+1182.208625489" observedRunningTime="2025-10-06 08:58:54.125721032 +0000 UTC m=+1184.915746612" watchObservedRunningTime="2025-10-06 08:58:54.126278698 +0000 UTC m=+1184.916304278" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.149510 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" podStartSLOduration=4.926037712 podStartE2EDuration="16.149489667s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.195148614 +0000 UTC m=+1170.985174184" lastFinishedPulling="2025-10-06 08:58:51.418600549 +0000 UTC m=+1182.208626139" observedRunningTime="2025-10-06 08:58:54.143932107 +0000 UTC m=+1184.933957687" watchObservedRunningTime="2025-10-06 08:58:54.149489667 +0000 UTC m=+1184.939515247" Oct 06 08:58:54 crc kubenswrapper[4989]: I1006 08:58:54.162456 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" podStartSLOduration=4.931089408 podStartE2EDuration="16.162437791s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.189908623 +0000 UTC m=+1170.979934203" lastFinishedPulling="2025-10-06 08:58:51.421257006 +0000 UTC m=+1182.211282586" observedRunningTime="2025-10-06 08:58:54.16206646 +0000 UTC m=+1184.952092070" watchObservedRunningTime="2025-10-06 08:58:54.162437791 +0000 UTC m=+1184.952463371" Oct 06 08:58:56 crc kubenswrapper[4989]: I1006 08:58:56.098758 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" event={"ID":"21ae0953-021f-40d1-aab0-2b693f29212a","Type":"ContainerStarted","Data":"29f5f98b6b2f0502764c8c55aab31b38ede5809854e294b702066d0903c86e8a"} Oct 06 08:58:56 crc kubenswrapper[4989]: I1006 08:58:56.099698 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:58:56 crc kubenswrapper[4989]: I1006 08:58:56.119709 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" podStartSLOduration=3.70871132 podStartE2EDuration="18.119685124s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.598010544 +0000 UTC m=+1171.388036134" lastFinishedPulling="2025-10-06 08:58:55.008984368 +0000 UTC m=+1185.799009938" observedRunningTime="2025-10-06 08:58:56.115410451 +0000 UTC m=+1186.905436041" watchObservedRunningTime="2025-10-06 08:58:56.119685124 +0000 UTC m=+1186.909710704" Oct 06 08:58:57 crc kubenswrapper[4989]: I1006 08:58:57.107820 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" event={"ID":"9d82b2b6-9374-4cbc-a72c-50aa000a0d52","Type":"ContainerStarted","Data":"ab4ef2811e2d39f9bd1abfacaa64aac35fa950b195f4e436d825b5ca9a9df818"} Oct 06 08:58:57 crc kubenswrapper[4989]: I1006 08:58:57.108303 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:58:57 crc kubenswrapper[4989]: I1006 08:58:57.122637 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" podStartSLOduration=3.331787859 podStartE2EDuration="19.122618533s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.680069261 +0000 UTC m=+1171.470094851" lastFinishedPulling="2025-10-06 08:58:56.470899945 +0000 UTC m=+1187.260925525" observedRunningTime="2025-10-06 08:58:57.120548733 +0000 UTC m=+1187.910574323" watchObservedRunningTime="2025-10-06 08:58:57.122618533 +0000 UTC m=+1187.912644113" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.122779 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" event={"ID":"428077cb-5a71-4f65-a37e-2384c4c0af19","Type":"ContainerStarted","Data":"cc4e19396fff84e17a12a476e2ea5823122621bb02389ee1fc59a58fa299e6ed"} Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.123795 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.127726 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" event={"ID":"9477ed03-7f88-4abf-bbcd-65a2888aff71","Type":"ContainerStarted","Data":"fb81b8fe8294340c41be45ea026a8b6852f4db2f7b2607e35e1ae776fed5e2b7"} Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.128204 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.141306 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" podStartSLOduration=2.95945239 podStartE2EDuration="20.141287825s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.629135392 +0000 UTC m=+1171.419160972" lastFinishedPulling="2025-10-06 08:58:57.810970827 +0000 UTC m=+1188.600996407" observedRunningTime="2025-10-06 08:58:58.141175522 +0000 UTC m=+1188.931201122" watchObservedRunningTime="2025-10-06 08:58:58.141287825 +0000 UTC m=+1188.931313395" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.161787 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" podStartSLOduration=2.993982356 podStartE2EDuration="20.161770446s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.635782784 +0000 UTC m=+1171.425808364" lastFinishedPulling="2025-10-06 08:58:57.803570874 +0000 UTC m=+1188.593596454" observedRunningTime="2025-10-06 08:58:58.158982066 +0000 UTC m=+1188.949007646" watchObservedRunningTime="2025-10-06 08:58:58.161770446 +0000 UTC m=+1188.951796026" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.481147 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-4skd5" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.503799 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6r4p2" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.541361 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-8h5sx" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.590154 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-25cxf" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.666248 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-phj9c" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.687349 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-dqjpg" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.721385 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-997ct" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.859457 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-mrz27" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.886380 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-79ksl" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.919707 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-m78db" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.959139 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-l62dg" Oct 06 08:58:58 crc kubenswrapper[4989]: I1006 08:58:58.991255 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fff4f" Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.134945 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" event={"ID":"522e1bb1-d22f-4256-b0f0-ae27447bd00a","Type":"ContainerStarted","Data":"f9336f052b4050f81dab2881c8bbc764c83b5b5b80fe1b2fa93035767db3d9a6"} Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.135123 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.136763 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" event={"ID":"3af9ba41-3325-4e75-a392-306b9999606b","Type":"ContainerStarted","Data":"a5aacd745478d237e170f8f68df682f8effb9fa8f59cb84e05933624623c4267"} Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.152973 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" podStartSLOduration=3.998270363 podStartE2EDuration="21.152957976s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.680275977 +0000 UTC m=+1171.470301557" lastFinishedPulling="2025-10-06 08:58:57.83496359 +0000 UTC m=+1188.624989170" observedRunningTime="2025-10-06 08:58:59.151263547 +0000 UTC m=+1189.941289127" watchObservedRunningTime="2025-10-06 08:58:59.152957976 +0000 UTC m=+1189.942983556" Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.180998 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" podStartSLOduration=3.640775772 podStartE2EDuration="21.180975244s" podCreationTimestamp="2025-10-06 08:58:38 +0000 UTC" firstStartedPulling="2025-10-06 08:58:40.949280546 +0000 UTC m=+1171.739306136" lastFinishedPulling="2025-10-06 08:58:58.489480028 +0000 UTC m=+1189.279505608" observedRunningTime="2025-10-06 08:58:59.177586326 +0000 UTC m=+1189.967611926" watchObservedRunningTime="2025-10-06 08:58:59.180975244 +0000 UTC m=+1189.971000824" Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.223321 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-nm6z6" Oct 06 08:58:59 crc kubenswrapper[4989]: I1006 08:58:59.341851 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-b5kt7" Oct 06 08:59:00 crc kubenswrapper[4989]: I1006 08:59:00.278362 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-cdzbs" Oct 06 08:59:00 crc kubenswrapper[4989]: I1006 08:59:00.367864 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:59:08 crc kubenswrapper[4989]: I1006 08:59:08.898086 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-gmfps" Oct 06 08:59:09 crc kubenswrapper[4989]: I1006 08:59:09.080178 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-jgjms" Oct 06 08:59:09 crc kubenswrapper[4989]: I1006 08:59:09.179373 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-lzjq2" Oct 06 08:59:09 crc kubenswrapper[4989]: I1006 08:59:09.210644 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kgbzk" Oct 06 08:59:09 crc kubenswrapper[4989]: I1006 08:59:09.380387 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-672zz" Oct 06 08:59:10 crc kubenswrapper[4989]: I1006 08:59:10.374867 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.392938 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.394532 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.397310 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.397476 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.398276 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.398539 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-fxj6h" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.404376 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.451321 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.451383 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfmsf\" (UniqueName: \"kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.466065 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.467521 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.469928 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.506857 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.552575 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvwzs\" (UniqueName: \"kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.552901 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.553023 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.553162 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfmsf\" (UniqueName: \"kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.553314 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.553932 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.575180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfmsf\" (UniqueName: \"kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf\") pod \"dnsmasq-dns-675f4bcbfc-wzq62\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.654814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.654903 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.654942 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvwzs\" (UniqueName: \"kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.655645 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.655776 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.672360 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvwzs\" (UniqueName: \"kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs\") pod \"dnsmasq-dns-78dd6ddcc-9twxw\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.710494 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:26 crc kubenswrapper[4989]: I1006 08:59:26.790717 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.140237 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.268228 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:27 crc kubenswrapper[4989]: W1006 08:59:27.277669 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod263c0634_c5d4_4e42_a2f9_f2d5e2791249.slice/crio-6ebf6c7a9ef3c3db0831b8968124573cf3068234edb7e79c0e8a7c5b09037661 WatchSource:0}: Error finding container 6ebf6c7a9ef3c3db0831b8968124573cf3068234edb7e79c0e8a7c5b09037661: Status 404 returned error can't find the container with id 6ebf6c7a9ef3c3db0831b8968124573cf3068234edb7e79c0e8a7c5b09037661 Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.343758 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" event={"ID":"8312ab70-d867-496f-aac6-8edf2ef13e2c","Type":"ContainerStarted","Data":"71cd9e36a41896c2dd845263eb4376047c6073b194453748a74f79349d53399a"} Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.346372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" event={"ID":"263c0634-c5d4-4e42-a2f9-f2d5e2791249","Type":"ContainerStarted","Data":"6ebf6c7a9ef3c3db0831b8968124573cf3068234edb7e79c0e8a7c5b09037661"} Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.497542 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.528688 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.530161 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.556481 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.569564 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.569704 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.569730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfqkn\" (UniqueName: \"kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.671252 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.671325 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.671351 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfqkn\" (UniqueName: \"kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.672383 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.672383 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.693437 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfqkn\" (UniqueName: \"kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn\") pod \"dnsmasq-dns-666b6646f7-ctmrv\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:27 crc kubenswrapper[4989]: I1006 08:59:27.848855 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.260886 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 08:59:28 crc kubenswrapper[4989]: W1006 08:59:28.276928 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b74bd43_f340_46e0_9c6d_db1b9ca7cbd1.slice/crio-f56f8ee890a4b5247755ae1176188050f33cc05fb62d4484debd0a98d82d4503 WatchSource:0}: Error finding container f56f8ee890a4b5247755ae1176188050f33cc05fb62d4484debd0a98d82d4503: Status 404 returned error can't find the container with id f56f8ee890a4b5247755ae1176188050f33cc05fb62d4484debd0a98d82d4503 Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.360729 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" event={"ID":"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1","Type":"ContainerStarted","Data":"f56f8ee890a4b5247755ae1176188050f33cc05fb62d4484debd0a98d82d4503"} Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.467953 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.506432 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.509704 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.536710 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.587112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.587203 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rvtm\" (UniqueName: \"kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.587238 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.673096 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.675567 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.677143 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-4dv7g" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683082 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683307 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683520 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683648 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683724 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683796 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.683889 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.688809 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.688872 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.688934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rvtm\" (UniqueName: \"kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.691255 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.691396 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.730375 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rvtm\" (UniqueName: \"kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm\") pod \"dnsmasq-dns-57d769cc4f-sth4d\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792165 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792295 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792357 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792406 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljp2w\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792496 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792569 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792693 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792726 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792769 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.792800 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.834969 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894620 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894702 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894766 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894840 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894882 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894913 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.894963 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljp2w\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.895014 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.895350 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.895754 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.896687 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.896859 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.898096 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.898290 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.898993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.900062 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.900347 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.901639 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.909557 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljp2w\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:28 crc kubenswrapper[4989]: I1006 08:59:28.927166 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " pod="openstack/rabbitmq-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.003065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.419376 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.547729 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.619180 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.621156 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.626978 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.627159 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.628927 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mshzm" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.629204 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.629350 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.629443 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.629537 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.638533 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709575 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709640 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnn66\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709681 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709856 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709896 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.709964 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.710008 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.710074 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.710104 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.710213 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.710255 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812276 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812343 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnn66\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812401 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812441 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812487 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812507 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812522 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812556 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.812886 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.815765 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.815988 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.816803 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.816981 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.819979 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.829023 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.829135 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.835759 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.835797 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnn66\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.835827 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.842445 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:29 crc kubenswrapper[4989]: I1006 08:59:29.953027 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 08:59:30 crc kubenswrapper[4989]: I1006 08:59:30.387242 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerStarted","Data":"9a93c9894dd0374e1b8060a5f0e19a9624a479c9f00dbaa15ad2c78f5b8a5f93"} Oct 06 08:59:30 crc kubenswrapper[4989]: I1006 08:59:30.389176 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" event={"ID":"db6cad76-cb75-4256-8a45-3eb27555e8a2","Type":"ContainerStarted","Data":"4ae0d0970c8f20c78614265e7d6e5d4ca1cca193f95dfddc4cf3b25dd92f4824"} Oct 06 08:59:30 crc kubenswrapper[4989]: I1006 08:59:30.553474 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.038424 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.040252 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.042353 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.043813 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.044184 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.044440 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.050773 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nbbn6" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.057044 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.057916 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134721 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134806 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134845 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134868 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134934 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134950 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.134972 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bp25\" (UniqueName: \"kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.135010 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.135034 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236230 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236269 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236332 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236348 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236371 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bp25\" (UniqueName: \"kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236393 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236435 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.236832 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.244160 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.245581 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.248949 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.249160 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.253425 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.255018 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.266988 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.270175 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bp25\" (UniqueName: \"kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.312508 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.386868 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.444627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerStarted","Data":"a6b77edea39c60d3ee36b124df90fe871d17d6826162373fdff14438b1068075"} Oct 06 08:59:31 crc kubenswrapper[4989]: I1006 08:59:31.935066 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.326377 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.327580 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.329529 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.329700 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.330636 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.332819 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xxt6b" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.345337 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.470683 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.470733 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.470882 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.470958 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.470976 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.471073 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.471121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.471198 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l26ml\" (UniqueName: \"kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.471236 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.572876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.572928 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.572964 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.572988 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.573005 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.573045 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.573066 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.573096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l26ml\" (UniqueName: \"kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.573119 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.576094 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.576103 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.576352 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.576513 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.577220 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.584359 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.585782 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.596516 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.608348 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.633533 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.642251 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l26ml\" (UniqueName: \"kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml\") pod \"openstack-cell1-galera-0\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.643551 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.647448 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.647630 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9xwkb" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.647762 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.657499 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.679690 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.777990 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.778259 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnf2q\" (UniqueName: \"kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.778295 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.778340 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.778359 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.879826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.879884 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnf2q\" (UniqueName: \"kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.879975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.880030 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.880049 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.880877 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.881630 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.883559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.900136 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:32 crc kubenswrapper[4989]: I1006 08:59:32.902415 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnf2q\" (UniqueName: \"kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q\") pod \"memcached-0\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " pod="openstack/memcached-0" Oct 06 08:59:33 crc kubenswrapper[4989]: I1006 08:59:33.006572 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 08:59:33 crc kubenswrapper[4989]: I1006 08:59:33.937048 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 08:59:33 crc kubenswrapper[4989]: I1006 08:59:33.937096 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.658257 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.659336 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.666720 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-99llx" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.668646 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.810914 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9mvw\" (UniqueName: \"kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw\") pod \"kube-state-metrics-0\" (UID: \"55218d49-6773-44cc-a7c0-38ac71f8f66d\") " pod="openstack/kube-state-metrics-0" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.915716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9mvw\" (UniqueName: \"kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw\") pod \"kube-state-metrics-0\" (UID: \"55218d49-6773-44cc-a7c0-38ac71f8f66d\") " pod="openstack/kube-state-metrics-0" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.941633 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9mvw\" (UniqueName: \"kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw\") pod \"kube-state-metrics-0\" (UID: \"55218d49-6773-44cc-a7c0-38ac71f8f66d\") " pod="openstack/kube-state-metrics-0" Oct 06 08:59:34 crc kubenswrapper[4989]: I1006 08:59:34.990033 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 08:59:35 crc kubenswrapper[4989]: I1006 08:59:35.518467 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerStarted","Data":"f5a8d3c2ce7e605df9caf12641cfdcf6f7ad14ac049fd896947dc9ee69cd638a"} Oct 06 08:59:36 crc kubenswrapper[4989]: I1006 08:59:36.470985 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.754940 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.765781 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.765804 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.768441 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-5n42f" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.768485 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.768687 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.768963 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.769190 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891136 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891277 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891312 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891346 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891376 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891405 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891434 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5kt4\" (UniqueName: \"kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.891464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993275 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993337 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993379 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993400 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.993438 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5kt4\" (UniqueName: \"kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.994898 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.995216 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.996922 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:38 crc kubenswrapper[4989]: I1006 08:59:38.997192 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.000737 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.001115 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.018374 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5kt4\" (UniqueName: \"kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.019732 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.033088 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.097149 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.446200 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.447477 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.454285 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.455035 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.455565 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jflvn" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.463567 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.474952 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.476767 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.490251 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605205 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605261 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkfbt\" (UniqueName: \"kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605290 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605335 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605358 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605390 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605411 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605427 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605461 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb2ts\" (UniqueName: \"kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605489 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605505 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605524 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.605543 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.706910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.706963 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707002 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707025 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb2ts\" (UniqueName: \"kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707122 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707144 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707168 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707225 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707257 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkfbt\" (UniqueName: \"kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707281 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707462 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707638 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.707642 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.708263 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.708379 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.708847 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.708902 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.711169 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.711331 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.712273 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.720200 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.726310 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb2ts\" (UniqueName: \"kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts\") pod \"ovn-controller-rzncg\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.726329 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkfbt\" (UniqueName: \"kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt\") pod \"ovn-controller-ovs-tmq59\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.777816 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg" Oct 06 08:59:39 crc kubenswrapper[4989]: I1006 08:59:39.803423 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:41 crc kubenswrapper[4989]: W1006 08:59:41.757850 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c674246_9c5c_4c8c_8d0b_360305a30818.slice/crio-afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7 WatchSource:0}: Error finding container afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7: Status 404 returned error can't find the container with id afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7 Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.179880 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.202199 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.210323 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.210727 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-bdnft" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.210938 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.210997 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.217528 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359376 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359441 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359535 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359592 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359634 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359670 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359696 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.359715 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sddhj\" (UniqueName: \"kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461568 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461641 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461701 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461725 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461756 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sddhj\" (UniqueName: \"kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461790 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461807 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.461950 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.468355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.468420 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.468495 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.469390 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.469847 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.469998 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.477763 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sddhj\" (UniqueName: \"kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.493921 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.528104 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 08:59:42 crc kubenswrapper[4989]: I1006 08:59:42.574451 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c674246-9c5c-4c8c-8d0b-360305a30818","Type":"ContainerStarted","Data":"afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7"} Oct 06 08:59:46 crc kubenswrapper[4989]: I1006 08:59:46.845586 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 08:59:47 crc kubenswrapper[4989]: W1006 08:59:47.211043 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d4f5ac8_6f42_41e9_91a9_dbe6a0a056a8.slice/crio-07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583 WatchSource:0}: Error finding container 07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583: Status 404 returned error can't find the container with id 07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583 Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.227987 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.228179 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pfmsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-wzq62_openstack(8312ab70-d867-496f-aac6-8edf2ef13e2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.229641 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" podUID="8312ab70-d867-496f-aac6-8edf2ef13e2c" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.258466 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.258947 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5rvtm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-sth4d_openstack(db6cad76-cb75-4256-8a45-3eb27555e8a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.260021 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.279611 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.279779 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfqkn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-ctmrv_openstack(5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.281257 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" Oct 06 08:59:47 crc kubenswrapper[4989]: I1006 08:59:47.374043 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 08:59:47 crc kubenswrapper[4989]: I1006 08:59:47.613530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerStarted","Data":"07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583"} Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.617275 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" Oct 06 08:59:47 crc kubenswrapper[4989]: E1006 08:59:47.617408 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" Oct 06 08:59:47 crc kubenswrapper[4989]: I1006 08:59:47.721076 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 08:59:51 crc kubenswrapper[4989]: W1006 08:59:51.677154 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b5f07d2_e2c4_406e_96e3_88e24d624a58.slice/crio-e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169 WatchSource:0}: Error finding container e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169: Status 404 returned error can't find the container with id e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169 Oct 06 08:59:51 crc kubenswrapper[4989]: E1006 08:59:51.758375 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 06 08:59:51 crc kubenswrapper[4989]: E1006 08:59:51.758949 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zvwzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-9twxw_openstack(263c0634-c5d4-4e42-a2f9-f2d5e2791249): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 08:59:51 crc kubenswrapper[4989]: E1006 08:59:51.762050 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" podUID="263c0634-c5d4-4e42-a2f9-f2d5e2791249" Oct 06 08:59:51 crc kubenswrapper[4989]: I1006 08:59:51.803876 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:51 crc kubenswrapper[4989]: I1006 08:59:51.920953 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfmsf\" (UniqueName: \"kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf\") pod \"8312ab70-d867-496f-aac6-8edf2ef13e2c\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " Oct 06 08:59:51 crc kubenswrapper[4989]: I1006 08:59:51.921405 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config\") pod \"8312ab70-d867-496f-aac6-8edf2ef13e2c\" (UID: \"8312ab70-d867-496f-aac6-8edf2ef13e2c\") " Oct 06 08:59:51 crc kubenswrapper[4989]: I1006 08:59:51.922116 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config" (OuterVolumeSpecName: "config") pod "8312ab70-d867-496f-aac6-8edf2ef13e2c" (UID: "8312ab70-d867-496f-aac6-8edf2ef13e2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:59:51 crc kubenswrapper[4989]: I1006 08:59:51.924729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf" (OuterVolumeSpecName: "kube-api-access-pfmsf") pod "8312ab70-d867-496f-aac6-8edf2ef13e2c" (UID: "8312ab70-d867-496f-aac6-8edf2ef13e2c"). InnerVolumeSpecName "kube-api-access-pfmsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.023792 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfmsf\" (UniqueName: \"kubernetes.io/projected/8312ab70-d867-496f-aac6-8edf2ef13e2c-kube-api-access-pfmsf\") on node \"crc\" DevicePath \"\"" Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.023919 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8312ab70-d867-496f-aac6-8edf2ef13e2c-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:59:52 crc kubenswrapper[4989]: W1006 08:59:52.299035 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39fb4b45_4966_4e65_b368_322d8c65e0c8.slice/crio-bfe39e3bdd8a5199ebb85e3e073a40a85088fc0bd159a7d2427941258c551a11 WatchSource:0}: Error finding container bfe39e3bdd8a5199ebb85e3e073a40a85088fc0bd159a7d2427941258c551a11: Status 404 returned error can't find the container with id bfe39e3bdd8a5199ebb85e3e073a40a85088fc0bd159a7d2427941258c551a11 Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.299728 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.376588 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 08:59:52 crc kubenswrapper[4989]: W1006 08:59:52.388812 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ec3e388_5714_48ab_86a5_01962c90587f.slice/crio-cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec WatchSource:0}: Error finding container cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec: Status 404 returned error can't find the container with id cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.468701 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.661791 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerStarted","Data":"cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.663677 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg" event={"ID":"6b5f07d2-e2c4-406e-96e3-88e24d624a58","Type":"ContainerStarted","Data":"e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.665167 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" event={"ID":"8312ab70-d867-496f-aac6-8edf2ef13e2c","Type":"ContainerDied","Data":"71cd9e36a41896c2dd845263eb4376047c6073b194453748a74f79349d53399a"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.665243 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-wzq62" Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.668164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c674246-9c5c-4c8c-8d0b-360305a30818","Type":"ContainerStarted","Data":"81180f30221310858c389627d3cef02db0be1d4733fdbc48816fac4ffa3f8a0f"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.669109 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.670954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerStarted","Data":"e37df7547070a50f2c8867346b9be9b45cff209265c630183e91bd4a1f02eeea"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.673124 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerStarted","Data":"6d1aa99512d591f42008e3253eb070ca752e10603a8f70402dc5c4cf6d4c9eaa"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.675770 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerStarted","Data":"e253d79428da6dc08fc7c77770484784d6ac2aeb7efd7e3493edb9854d9065e6"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.680184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55218d49-6773-44cc-a7c0-38ac71f8f66d","Type":"ContainerStarted","Data":"9be20d61f36b4f7d6aa2aec9cfbd9c61805a60775e082b38cfcb0fbcadcf56f2"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.683437 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerStarted","Data":"bfe39e3bdd8a5199ebb85e3e073a40a85088fc0bd159a7d2427941258c551a11"} Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.717857 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.723957 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-wzq62"] Oct 06 08:59:52 crc kubenswrapper[4989]: I1006 08:59:52.774421 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.729801059 podStartE2EDuration="20.774405263s" podCreationTimestamp="2025-10-06 08:59:32 +0000 UTC" firstStartedPulling="2025-10-06 08:59:41.780720065 +0000 UTC m=+1232.570745655" lastFinishedPulling="2025-10-06 08:59:51.825324289 +0000 UTC m=+1242.615349859" observedRunningTime="2025-10-06 08:59:52.738277364 +0000 UTC m=+1243.528302944" watchObservedRunningTime="2025-10-06 08:59:52.774405263 +0000 UTC m=+1243.564430843" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.366818 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.451903 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc\") pod \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.452009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvwzs\" (UniqueName: \"kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs\") pod \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.452099 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config\") pod \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\" (UID: \"263c0634-c5d4-4e42-a2f9-f2d5e2791249\") " Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.453254 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config" (OuterVolumeSpecName: "config") pod "263c0634-c5d4-4e42-a2f9-f2d5e2791249" (UID: "263c0634-c5d4-4e42-a2f9-f2d5e2791249"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.454400 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "263c0634-c5d4-4e42-a2f9-f2d5e2791249" (UID: "263c0634-c5d4-4e42-a2f9-f2d5e2791249"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.466516 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs" (OuterVolumeSpecName: "kube-api-access-zvwzs") pod "263c0634-c5d4-4e42-a2f9-f2d5e2791249" (UID: "263c0634-c5d4-4e42-a2f9-f2d5e2791249"). InnerVolumeSpecName "kube-api-access-zvwzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.554887 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-config\") on node \"crc\" DevicePath \"\"" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.554923 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/263c0634-c5d4-4e42-a2f9-f2d5e2791249-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.554936 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvwzs\" (UniqueName: \"kubernetes.io/projected/263c0634-c5d4-4e42-a2f9-f2d5e2791249-kube-api-access-zvwzs\") on node \"crc\" DevicePath \"\"" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.693559 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" event={"ID":"263c0634-c5d4-4e42-a2f9-f2d5e2791249","Type":"ContainerDied","Data":"6ebf6c7a9ef3c3db0831b8968124573cf3068234edb7e79c0e8a7c5b09037661"} Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.693579 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9twxw" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.695372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerStarted","Data":"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4"} Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.697599 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerStarted","Data":"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b"} Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.801733 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.807525 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9twxw"] Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.946121 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="263c0634-c5d4-4e42-a2f9-f2d5e2791249" path="/var/lib/kubelet/pods/263c0634-c5d4-4e42-a2f9-f2d5e2791249/volumes" Oct 06 08:59:53 crc kubenswrapper[4989]: I1006 08:59:53.946495 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8312ab70-d867-496f-aac6-8edf2ef13e2c" path="/var/lib/kubelet/pods/8312ab70-d867-496f-aac6-8edf2ef13e2c/volumes" Oct 06 08:59:55 crc kubenswrapper[4989]: I1006 08:59:55.711673 4989 generic.go:334] "Generic (PLEG): container finished" podID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerID="6d1aa99512d591f42008e3253eb070ca752e10603a8f70402dc5c4cf6d4c9eaa" exitCode=0 Oct 06 08:59:55 crc kubenswrapper[4989]: I1006 08:59:55.711762 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerDied","Data":"6d1aa99512d591f42008e3253eb070ca752e10603a8f70402dc5c4cf6d4c9eaa"} Oct 06 08:59:55 crc kubenswrapper[4989]: I1006 08:59:55.713907 4989 generic.go:334] "Generic (PLEG): container finished" podID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerID="e253d79428da6dc08fc7c77770484784d6ac2aeb7efd7e3493edb9854d9065e6" exitCode=0 Oct 06 08:59:55 crc kubenswrapper[4989]: I1006 08:59:55.713931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerDied","Data":"e253d79428da6dc08fc7c77770484784d6ac2aeb7efd7e3493edb9854d9065e6"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.722281 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerStarted","Data":"db6a98c9f7876b97505eaf206617e7f0a26dba08f6ba38b5d990e6763b2bee16"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.723757 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerStarted","Data":"888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.725423 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerStarted","Data":"c0e57d5d18f141af4479840759a41e612d6a1d950914d8c0d137ed2381bc7ce1"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.727382 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55218d49-6773-44cc-a7c0-38ac71f8f66d","Type":"ContainerStarted","Data":"354435a487e4639127dcd86ea2b36527b64d79471e09035787232d88307f80f1"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.727454 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.729123 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerStarted","Data":"c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.732108 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg" event={"ID":"6b5f07d2-e2c4-406e-96e3-88e24d624a58","Type":"ContainerStarted","Data":"23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.732258 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rzncg" Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.733883 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerStarted","Data":"77075241e0c723fffbcdd5be38b37be16523bbbe9459476084190610319b44be"} Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.754983 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=18.027029014 podStartE2EDuration="22.75496164s" podCreationTimestamp="2025-10-06 08:59:34 +0000 UTC" firstStartedPulling="2025-10-06 08:59:51.683246911 +0000 UTC m=+1242.473272491" lastFinishedPulling="2025-10-06 08:59:56.411179537 +0000 UTC m=+1247.201205117" observedRunningTime="2025-10-06 08:59:56.753771315 +0000 UTC m=+1247.543796895" watchObservedRunningTime="2025-10-06 08:59:56.75496164 +0000 UTC m=+1247.544987230" Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.778550 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.778528608 podStartE2EDuration="25.778528608s" podCreationTimestamp="2025-10-06 08:59:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 08:59:56.772067962 +0000 UTC m=+1247.562093542" watchObservedRunningTime="2025-10-06 08:59:56.778528608 +0000 UTC m=+1247.568554198" Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.792799 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rzncg" podStartSLOduration=13.064895823 podStartE2EDuration="17.792775258s" podCreationTimestamp="2025-10-06 08:59:39 +0000 UTC" firstStartedPulling="2025-10-06 08:59:51.683258521 +0000 UTC m=+1242.473284111" lastFinishedPulling="2025-10-06 08:59:56.411137966 +0000 UTC m=+1247.201163546" observedRunningTime="2025-10-06 08:59:56.787774454 +0000 UTC m=+1247.577800044" watchObservedRunningTime="2025-10-06 08:59:56.792775258 +0000 UTC m=+1247.582800838" Oct 06 08:59:56 crc kubenswrapper[4989]: I1006 08:59:56.810259 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.558847469 podStartE2EDuration="26.81023569s" podCreationTimestamp="2025-10-06 08:59:30 +0000 UTC" firstStartedPulling="2025-10-06 08:59:35.441685202 +0000 UTC m=+1226.231710782" lastFinishedPulling="2025-10-06 08:59:51.693073423 +0000 UTC m=+1242.483099003" observedRunningTime="2025-10-06 08:59:56.807625225 +0000 UTC m=+1247.597650805" watchObservedRunningTime="2025-10-06 08:59:56.81023569 +0000 UTC m=+1247.600261270" Oct 06 08:59:57 crc kubenswrapper[4989]: I1006 08:59:57.744835 4989 generic.go:334] "Generic (PLEG): container finished" podID="29637231-ec41-4e57-af1f-18046cc06cf6" containerID="db6a98c9f7876b97505eaf206617e7f0a26dba08f6ba38b5d990e6763b2bee16" exitCode=0 Oct 06 08:59:57 crc kubenswrapper[4989]: I1006 08:59:57.744943 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerDied","Data":"db6a98c9f7876b97505eaf206617e7f0a26dba08f6ba38b5d990e6763b2bee16"} Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.009457 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.754639 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerStarted","Data":"9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825"} Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.754953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerStarted","Data":"8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de"} Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.754980 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.754991 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 08:59:58 crc kubenswrapper[4989]: I1006 08:59:58.774966 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-tmq59" podStartSLOduration=15.953768683 podStartE2EDuration="19.774949834s" podCreationTimestamp="2025-10-06 08:59:39 +0000 UTC" firstStartedPulling="2025-10-06 08:59:52.590557872 +0000 UTC m=+1243.380583452" lastFinishedPulling="2025-10-06 08:59:56.411739023 +0000 UTC m=+1247.201764603" observedRunningTime="2025-10-06 08:59:58.770898937 +0000 UTC m=+1249.560924517" watchObservedRunningTime="2025-10-06 08:59:58.774949834 +0000 UTC m=+1249.564975414" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.132255 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv"] Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.135918 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.139795 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.140056 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.145363 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv"] Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.178203 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.178401 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrrqt\" (UniqueName: \"kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.178435 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.280616 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.280762 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrrqt\" (UniqueName: \"kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.280790 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.281794 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.285626 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.298077 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrrqt\" (UniqueName: \"kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt\") pod \"collect-profiles-29329020-dd9dv\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.481955 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.784089 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerStarted","Data":"a2d8764d48f0504acb3368e626f02f443813a1caafd1fee9a37fbf6015805564"} Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.786115 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerStarted","Data":"50a79f09910e9fff54369d71bc6811a462f4e436d855283ce8d406341063c274"} Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.824790 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=16.123660873 podStartE2EDuration="23.824765245s" podCreationTimestamp="2025-10-06 08:59:37 +0000 UTC" firstStartedPulling="2025-10-06 08:59:52.301171714 +0000 UTC m=+1243.091197304" lastFinishedPulling="2025-10-06 09:00:00.002276096 +0000 UTC m=+1250.792301676" observedRunningTime="2025-10-06 09:00:00.818270438 +0000 UTC m=+1251.608296018" watchObservedRunningTime="2025-10-06 09:00:00.824765245 +0000 UTC m=+1251.614790825" Oct 06 09:00:00 crc kubenswrapper[4989]: I1006 09:00:00.849641 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.246589652 podStartE2EDuration="19.849624441s" podCreationTimestamp="2025-10-06 08:59:41 +0000 UTC" firstStartedPulling="2025-10-06 08:59:52.39207439 +0000 UTC m=+1243.182099970" lastFinishedPulling="2025-10-06 08:59:59.995109179 +0000 UTC m=+1250.785134759" observedRunningTime="2025-10-06 09:00:00.847568972 +0000 UTC m=+1251.637594562" watchObservedRunningTime="2025-10-06 09:00:00.849624441 +0000 UTC m=+1251.639650021" Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.076027 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv"] Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.387662 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.387720 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.436660 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.795146 4989 generic.go:334] "Generic (PLEG): container finished" podID="7fbc8169-114f-43b8-a6b2-671aa3327536" containerID="78e06aa2ac711ad850e7a81bd66a5aca2cac653591e37746d8313f6504eadb88" exitCode=0 Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.795197 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" event={"ID":"7fbc8169-114f-43b8-a6b2-671aa3327536","Type":"ContainerDied","Data":"78e06aa2ac711ad850e7a81bd66a5aca2cac653591e37746d8313f6504eadb88"} Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.795474 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" event={"ID":"7fbc8169-114f-43b8-a6b2-671aa3327536","Type":"ContainerStarted","Data":"03c3439160763e3100476d6db0b3d218f4cfbcdb4b77546ede955bdba116b656"} Oct 06 09:00:01 crc kubenswrapper[4989]: I1006 09:00:01.850097 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.528854 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.643384 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jbcbt"] Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.644418 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.655798 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbcbt"] Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.680837 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.681439 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.730850 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.731051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tqhn\" (UniqueName: \"kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn\") pod \"keystone-db-create-jbcbt\" (UID: \"0f04a1cc-7dd1-4987-bbb1-258b06fc369f\") " pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.804303 4989 generic.go:334] "Generic (PLEG): container finished" podID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerID="aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd" exitCode=0 Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.804390 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" event={"ID":"db6cad76-cb75-4256-8a45-3eb27555e8a2","Type":"ContainerDied","Data":"aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd"} Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.806073 4989 generic.go:334] "Generic (PLEG): container finished" podID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerID="36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5" exitCode=0 Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.806235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" event={"ID":"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1","Type":"ContainerDied","Data":"36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5"} Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.832244 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tqhn\" (UniqueName: \"kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn\") pod \"keystone-db-create-jbcbt\" (UID: \"0f04a1cc-7dd1-4987-bbb1-258b06fc369f\") " pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.850587 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tqhn\" (UniqueName: \"kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn\") pod \"keystone-db-create-jbcbt\" (UID: \"0f04a1cc-7dd1-4987-bbb1-258b06fc369f\") " pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.872875 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.899606 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nf448"] Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.901066 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nf448" Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.911240 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nf448"] Oct 06 09:00:02 crc kubenswrapper[4989]: I1006 09:00:02.961722 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.038397 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rt5v\" (UniqueName: \"kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v\") pod \"placement-db-create-nf448\" (UID: \"73589ab3-b75d-4d92-b312-2464ba72317a\") " pod="openstack/placement-db-create-nf448" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.098190 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.143629 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rt5v\" (UniqueName: \"kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v\") pod \"placement-db-create-nf448\" (UID: \"73589ab3-b75d-4d92-b312-2464ba72317a\") " pod="openstack/placement-db-create-nf448" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.159901 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.167789 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rt5v\" (UniqueName: \"kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v\") pod \"placement-db-create-nf448\" (UID: \"73589ab3-b75d-4d92-b312-2464ba72317a\") " pod="openstack/placement-db-create-nf448" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.197794 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.246240 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrrqt\" (UniqueName: \"kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt\") pod \"7fbc8169-114f-43b8-a6b2-671aa3327536\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.246289 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume\") pod \"7fbc8169-114f-43b8-a6b2-671aa3327536\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.246379 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume\") pod \"7fbc8169-114f-43b8-a6b2-671aa3327536\" (UID: \"7fbc8169-114f-43b8-a6b2-671aa3327536\") " Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.247395 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume" (OuterVolumeSpecName: "config-volume") pod "7fbc8169-114f-43b8-a6b2-671aa3327536" (UID: "7fbc8169-114f-43b8-a6b2-671aa3327536"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.250905 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt" (OuterVolumeSpecName: "kube-api-access-qrrqt") pod "7fbc8169-114f-43b8-a6b2-671aa3327536" (UID: "7fbc8169-114f-43b8-a6b2-671aa3327536"). InnerVolumeSpecName "kube-api-access-qrrqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.253497 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7fbc8169-114f-43b8-a6b2-671aa3327536" (UID: "7fbc8169-114f-43b8-a6b2-671aa3327536"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.288427 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nf448" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.347828 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrrqt\" (UniqueName: \"kubernetes.io/projected/7fbc8169-114f-43b8-a6b2-671aa3327536-kube-api-access-qrrqt\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.347859 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fbc8169-114f-43b8-a6b2-671aa3327536-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.347870 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fbc8169-114f-43b8-a6b2-671aa3327536-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.444916 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbcbt"] Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.529255 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.575079 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.705847 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nf448"] Oct 06 09:00:03 crc kubenswrapper[4989]: W1006 09:00:03.708019 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73589ab3_b75d_4d92_b312_2464ba72317a.slice/crio-20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6 WatchSource:0}: Error finding container 20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6: Status 404 returned error can't find the container with id 20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6 Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.845169 4989 generic.go:334] "Generic (PLEG): container finished" podID="0f04a1cc-7dd1-4987-bbb1-258b06fc369f" containerID="5e27806c6c9bbfa7148ab8855e4f5365764ef0a9865db06ea4f1be2251cbccb8" exitCode=0 Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.845270 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbcbt" event={"ID":"0f04a1cc-7dd1-4987-bbb1-258b06fc369f","Type":"ContainerDied","Data":"5e27806c6c9bbfa7148ab8855e4f5365764ef0a9865db06ea4f1be2251cbccb8"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.845300 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbcbt" event={"ID":"0f04a1cc-7dd1-4987-bbb1-258b06fc369f","Type":"ContainerStarted","Data":"b37bc8a097ae6c7d09561447d0a964bda1b9ff6fb605694b9ac16a63d8fa6817"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.850840 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nf448" event={"ID":"73589ab3-b75d-4d92-b312-2464ba72317a","Type":"ContainerStarted","Data":"20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.869051 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" event={"ID":"db6cad76-cb75-4256-8a45-3eb27555e8a2","Type":"ContainerStarted","Data":"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.869393 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.894215 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" event={"ID":"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1","Type":"ContainerStarted","Data":"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.894929 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.897936 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.902323 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv" event={"ID":"7fbc8169-114f-43b8-a6b2-671aa3327536","Type":"ContainerDied","Data":"03c3439160763e3100476d6db0b3d218f4cfbcdb4b77546ede955bdba116b656"} Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.902352 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03c3439160763e3100476d6db0b3d218f4cfbcdb4b77546ede955bdba116b656" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.903125 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.903124 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" podStartSLOduration=-9223372000.951674 podStartE2EDuration="35.903102977s" podCreationTimestamp="2025-10-06 08:59:28 +0000 UTC" firstStartedPulling="2025-10-06 08:59:29.434169644 +0000 UTC m=+1220.224195224" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:03.901440789 +0000 UTC m=+1254.691466369" watchObservedRunningTime="2025-10-06 09:00:03.903102977 +0000 UTC m=+1254.693128557" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.935367 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.935678 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.937117 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" podStartSLOduration=3.538613464 podStartE2EDuration="36.937092595s" podCreationTimestamp="2025-10-06 08:59:27 +0000 UTC" firstStartedPulling="2025-10-06 08:59:28.285362518 +0000 UTC m=+1219.075388098" lastFinishedPulling="2025-10-06 09:00:01.683841649 +0000 UTC m=+1252.473867229" observedRunningTime="2025-10-06 09:00:03.929775685 +0000 UTC m=+1254.719801275" watchObservedRunningTime="2025-10-06 09:00:03.937092595 +0000 UTC m=+1254.727118195" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.965931 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 06 09:00:03 crc kubenswrapper[4989]: I1006 09:00:03.971348 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.224972 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.259855 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:04 crc kubenswrapper[4989]: E1006 09:00:04.260477 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbc8169-114f-43b8-a6b2-671aa3327536" containerName="collect-profiles" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.260575 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbc8169-114f-43b8-a6b2-671aa3327536" containerName="collect-profiles" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.260890 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbc8169-114f-43b8-a6b2-671aa3327536" containerName="collect-profiles" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.262027 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.265432 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.278642 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.318300 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.319439 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.322082 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.331572 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.366025 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z45xv\" (UniqueName: \"kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.366109 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.366161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.366285 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.420407 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.421899 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.424294 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-cxwvg" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.424673 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.428024 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.436244 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.440086 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.447132 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.467918 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z45xv\" (UniqueName: \"kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.467991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468039 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468072 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468102 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468189 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468222 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468242 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.468268 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcr4q\" (UniqueName: \"kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.469504 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.470284 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.476917 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.493395 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.497287 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.504055 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.504493 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z45xv\" (UniqueName: \"kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv\") pod \"dnsmasq-dns-7f896c8c65-9wbnx\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.529076 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.569753 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.569910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.569954 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.569977 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570006 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570028 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570056 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htm69\" (UniqueName: \"kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570077 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570167 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570210 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570234 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570263 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcr4q\" (UniqueName: \"kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570293 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570636 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.570643 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.572303 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.573845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.574775 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.592750 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcr4q\" (UniqueName: \"kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q\") pod \"ovn-controller-metrics-spqbp\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.595619 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.647636 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672301 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672397 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672420 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672459 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672504 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672675 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htm69\" (UniqueName: \"kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672729 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672765 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672786 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pc8\" (UniqueName: \"kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672937 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.672975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.673451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.673514 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.673742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.675815 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.675851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.676896 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.696227 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htm69\" (UniqueName: \"kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69\") pod \"ovn-northd-0\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.745567 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.774571 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.774738 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.774777 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.774824 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.774844 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pc8\" (UniqueName: \"kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.775808 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.775817 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.775980 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.776397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.793462 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pc8\" (UniqueName: \"kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8\") pod \"dnsmasq-dns-86db49b7ff-qg47h\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.853325 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.950582 4989 generic.go:334] "Generic (PLEG): container finished" podID="73589ab3-b75d-4d92-b312-2464ba72317a" containerID="a2ae3fd46babdad278c9956c887f5c124a9304e8a4b2bcc60897796e72e63572" exitCode=0 Oct 06 09:00:04 crc kubenswrapper[4989]: I1006 09:00:04.950773 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nf448" event={"ID":"73589ab3-b75d-4d92-b312-2464ba72317a","Type":"ContainerDied","Data":"a2ae3fd46babdad278c9956c887f5c124a9304e8a4b2bcc60897796e72e63572"} Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:04.999959 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.096617 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.202800 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.212811 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.242473 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:05 crc kubenswrapper[4989]: W1006 09:00:05.266429 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2445aa06_1e2f_41fe_ae4d_a998f7874955.slice/crio-cbb6d3fbedb7c7e24b23449713d9d60e1f50002adb80d74c5257ca74cf66e50d WatchSource:0}: Error finding container cbb6d3fbedb7c7e24b23449713d9d60e1f50002adb80d74c5257ca74cf66e50d: Status 404 returned error can't find the container with id cbb6d3fbedb7c7e24b23449713d9d60e1f50002adb80d74c5257ca74cf66e50d Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.271706 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.281278 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.297364 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.297404 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.297497 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9nnb\" (UniqueName: \"kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.297532 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.297563 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.365077 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.400168 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.400308 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.400339 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.400463 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9nnb\" (UniqueName: \"kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.400528 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.401858 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.402148 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.402421 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.404117 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.427124 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9nnb\" (UniqueName: \"kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb\") pod \"dnsmasq-dns-698758b865-zzclq\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.605770 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:05 crc kubenswrapper[4989]: W1006 09:00:05.626561 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a76137_a1aa_43ca_8c76_26aef58596ae.slice/crio-71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81 WatchSource:0}: Error finding container 71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81: Status 404 returned error can't find the container with id 71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81 Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.627739 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.705376 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tqhn\" (UniqueName: \"kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn\") pod \"0f04a1cc-7dd1-4987-bbb1-258b06fc369f\" (UID: \"0f04a1cc-7dd1-4987-bbb1-258b06fc369f\") " Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.709911 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn" (OuterVolumeSpecName: "kube-api-access-5tqhn") pod "0f04a1cc-7dd1-4987-bbb1-258b06fc369f" (UID: "0f04a1cc-7dd1-4987-bbb1-258b06fc369f"). InnerVolumeSpecName "kube-api-access-5tqhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.738028 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:05 crc kubenswrapper[4989]: I1006 09:00:05.807937 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tqhn\" (UniqueName: \"kubernetes.io/projected/0f04a1cc-7dd1-4987-bbb1-258b06fc369f-kube-api-access-5tqhn\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.003985 4989 generic.go:334] "Generic (PLEG): container finished" podID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerID="839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9" exitCode=0 Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.004043 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" event={"ID":"10a76137-a1aa-43ca-8c76-26aef58596ae","Type":"ContainerDied","Data":"839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.004067 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" event={"ID":"10a76137-a1aa-43ca-8c76-26aef58596ae","Type":"ContainerStarted","Data":"71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.018881 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerStarted","Data":"2a942f4481e91b5f1414b133e3b415ffe137c9225e1a7712198ba31b70f75e61"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.043915 4989 generic.go:334] "Generic (PLEG): container finished" podID="82778b9a-f721-4dce-ad71-9d36af2248ae" containerID="ffdb346adf2d6a16ff638180670a9c70883f99f9c7546b731a78608c9c33cfa6" exitCode=0 Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.044001 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" event={"ID":"82778b9a-f721-4dce-ad71-9d36af2248ae","Type":"ContainerDied","Data":"ffdb346adf2d6a16ff638180670a9c70883f99f9c7546b731a78608c9c33cfa6"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.044025 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" event={"ID":"82778b9a-f721-4dce-ad71-9d36af2248ae","Type":"ContainerStarted","Data":"5b19a6c44a5ac8256870edb2ad2d1a789425b36f9112a06f89352bcbb6d014ef"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.057388 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-spqbp" event={"ID":"2445aa06-1e2f-41fe-ae4d-a998f7874955","Type":"ContainerStarted","Data":"d527d1e974bd70387c4d59349474b57da4a2251238da24621c65f48e2aba78db"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.057431 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-spqbp" event={"ID":"2445aa06-1e2f-41fe-ae4d-a998f7874955","Type":"ContainerStarted","Data":"cbb6d3fbedb7c7e24b23449713d9d60e1f50002adb80d74c5257ca74cf66e50d"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.076114 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="dnsmasq-dns" containerID="cri-o://bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962" gracePeriod=10 Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.076460 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbcbt" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.077172 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbcbt" event={"ID":"0f04a1cc-7dd1-4987-bbb1-258b06fc369f","Type":"ContainerDied","Data":"b37bc8a097ae6c7d09561447d0a964bda1b9ff6fb605694b9ac16a63d8fa6817"} Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.077201 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b37bc8a097ae6c7d09561447d0a964bda1b9ff6fb605694b9ac16a63d8fa6817" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.078016 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="dnsmasq-dns" containerID="cri-o://2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278" gracePeriod=10 Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.141402 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-spqbp" podStartSLOduration=2.141385083 podStartE2EDuration="2.141385083s" podCreationTimestamp="2025-10-06 09:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:06.132395575 +0000 UTC m=+1256.922421155" watchObservedRunningTime="2025-10-06 09:00:06.141385083 +0000 UTC m=+1256.931410663" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.173981 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:06 crc kubenswrapper[4989]: W1006 09:00:06.184436 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f360dc_50e0_4b88_b25f_cc838c3a95f4.slice/crio-26321d24117b92712d3bbf3353a55f35166412d8bfecb155455cb49ea24b5869 WatchSource:0}: Error finding container 26321d24117b92712d3bbf3353a55f35166412d8bfecb155455cb49ea24b5869: Status 404 returned error can't find the container with id 26321d24117b92712d3bbf3353a55f35166412d8bfecb155455cb49ea24b5869 Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.228092 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.228479 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f04a1cc-7dd1-4987-bbb1-258b06fc369f" containerName="mariadb-database-create" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.228493 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f04a1cc-7dd1-4987-bbb1-258b06fc369f" containerName="mariadb-database-create" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.228744 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f04a1cc-7dd1-4987-bbb1-258b06fc369f" containerName="mariadb-database-create" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.254399 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.259375 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-qszr8" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.259577 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.259714 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.259805 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.271096 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.420643 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.421014 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.421069 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.421091 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z5xf\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.421158 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.508019 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.524373 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.524447 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.524504 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.524529 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z5xf\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.524577 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.525320 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.525452 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.525476 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.525522 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:00:07.025502848 +0000 UTC m=+1257.815528428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : configmap "swift-ring-files" not found Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.526270 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.526569 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.558354 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z5xf\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.587967 4989 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 06 09:00:06 crc kubenswrapper[4989]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/10a76137-a1aa-43ca-8c76-26aef58596ae/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 06 09:00:06 crc kubenswrapper[4989]: > podSandboxID="71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81" Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.588250 4989 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 06 09:00:06 crc kubenswrapper[4989]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n599h5cbh7ch5d4h66fh676hdbh546h95h88h5ffh55ch7fhch57ch687hddhc7h5fdh57dh674h56fh64ch98h9bh557h55dh646h54ch54fh5c4h597q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r7pc8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-86db49b7ff-qg47h_openstack(10a76137-a1aa-43ca-8c76-26aef58596ae): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/10a76137-a1aa-43ca-8c76-26aef58596ae/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 06 09:00:06 crc kubenswrapper[4989]: > logger="UnhandledError" Oct 06 09:00:06 crc kubenswrapper[4989]: E1006 09:00:06.589513 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/10a76137-a1aa-43ca-8c76-26aef58596ae/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.600704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.626158 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb\") pod \"82778b9a-f721-4dce-ad71-9d36af2248ae\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.626221 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config\") pod \"82778b9a-f721-4dce-ad71-9d36af2248ae\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.626266 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc\") pod \"82778b9a-f721-4dce-ad71-9d36af2248ae\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.626298 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z45xv\" (UniqueName: \"kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv\") pod \"82778b9a-f721-4dce-ad71-9d36af2248ae\" (UID: \"82778b9a-f721-4dce-ad71-9d36af2248ae\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.630768 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv" (OuterVolumeSpecName: "kube-api-access-z45xv") pod "82778b9a-f721-4dce-ad71-9d36af2248ae" (UID: "82778b9a-f721-4dce-ad71-9d36af2248ae"). InnerVolumeSpecName "kube-api-access-z45xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.647971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config" (OuterVolumeSpecName: "config") pod "82778b9a-f721-4dce-ad71-9d36af2248ae" (UID: "82778b9a-f721-4dce-ad71-9d36af2248ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.665179 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82778b9a-f721-4dce-ad71-9d36af2248ae" (UID: "82778b9a-f721-4dce-ad71-9d36af2248ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.669461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82778b9a-f721-4dce-ad71-9d36af2248ae" (UID: "82778b9a-f721-4dce-ad71-9d36af2248ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.671126 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nf448" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.727831 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.727870 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z45xv\" (UniqueName: \"kubernetes.io/projected/82778b9a-f721-4dce-ad71-9d36af2248ae-kube-api-access-z45xv\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.727881 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.727889 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82778b9a-f721-4dce-ad71-9d36af2248ae-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.749580 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.754930 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828349 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc\") pod \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rt5v\" (UniqueName: \"kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v\") pod \"73589ab3-b75d-4d92-b312-2464ba72317a\" (UID: \"73589ab3-b75d-4d92-b312-2464ba72317a\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828511 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rvtm\" (UniqueName: \"kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm\") pod \"db6cad76-cb75-4256-8a45-3eb27555e8a2\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config\") pod \"db6cad76-cb75-4256-8a45-3eb27555e8a2\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828686 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc\") pod \"db6cad76-cb75-4256-8a45-3eb27555e8a2\" (UID: \"db6cad76-cb75-4256-8a45-3eb27555e8a2\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828722 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config\") pod \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.828761 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfqkn\" (UniqueName: \"kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn\") pod \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\" (UID: \"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1\") " Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.836516 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn" (OuterVolumeSpecName: "kube-api-access-rfqkn") pod "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" (UID: "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1"). InnerVolumeSpecName "kube-api-access-rfqkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.839345 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm" (OuterVolumeSpecName: "kube-api-access-5rvtm") pod "db6cad76-cb75-4256-8a45-3eb27555e8a2" (UID: "db6cad76-cb75-4256-8a45-3eb27555e8a2"). InnerVolumeSpecName "kube-api-access-5rvtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.839938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v" (OuterVolumeSpecName: "kube-api-access-7rt5v") pod "73589ab3-b75d-4d92-b312-2464ba72317a" (UID: "73589ab3-b75d-4d92-b312-2464ba72317a"). InnerVolumeSpecName "kube-api-access-7rt5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.875739 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config" (OuterVolumeSpecName: "config") pod "db6cad76-cb75-4256-8a45-3eb27555e8a2" (UID: "db6cad76-cb75-4256-8a45-3eb27555e8a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.886317 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" (UID: "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.888197 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config" (OuterVolumeSpecName: "config") pod "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" (UID: "5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.888689 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db6cad76-cb75-4256-8a45-3eb27555e8a2" (UID: "db6cad76-cb75-4256-8a45-3eb27555e8a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930744 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930785 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rt5v\" (UniqueName: \"kubernetes.io/projected/73589ab3-b75d-4d92-b312-2464ba72317a-kube-api-access-7rt5v\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930801 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rvtm\" (UniqueName: \"kubernetes.io/projected/db6cad76-cb75-4256-8a45-3eb27555e8a2-kube-api-access-5rvtm\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930834 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930847 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db6cad76-cb75-4256-8a45-3eb27555e8a2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930859 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:06 crc kubenswrapper[4989]: I1006 09:00:06.930868 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfqkn\" (UniqueName: \"kubernetes.io/projected/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1-kube-api-access-rfqkn\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.032960 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.033094 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.033327 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.033382 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:00:08.033364384 +0000 UTC m=+1258.823390024 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : configmap "swift-ring-files" not found Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.087061 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nf448" event={"ID":"73589ab3-b75d-4d92-b312-2464ba72317a","Type":"ContainerDied","Data":"20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.087092 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20027618f96d5300062861c95663055f382dc0bedd9e81f884ecf8ac563e80f6" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.087093 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nf448" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.088601 4989 generic.go:334] "Generic (PLEG): container finished" podID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerID="6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad" exitCode=0 Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.088640 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zzclq" event={"ID":"95f360dc-50e0-4b88-b25f-cc838c3a95f4","Type":"ContainerDied","Data":"6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.088669 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zzclq" event={"ID":"95f360dc-50e0-4b88-b25f-cc838c3a95f4","Type":"ContainerStarted","Data":"26321d24117b92712d3bbf3353a55f35166412d8bfecb155455cb49ea24b5869"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.101912 4989 generic.go:334] "Generic (PLEG): container finished" podID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerID="bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962" exitCode=0 Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.102068 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.102360 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" event={"ID":"db6cad76-cb75-4256-8a45-3eb27555e8a2","Type":"ContainerDied","Data":"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.102451 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-sth4d" event={"ID":"db6cad76-cb75-4256-8a45-3eb27555e8a2","Type":"ContainerDied","Data":"4ae0d0970c8f20c78614265e7d6e5d4ca1cca193f95dfddc4cf3b25dd92f4824"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.102471 4989 scope.go:117] "RemoveContainer" containerID="bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.116376 4989 generic.go:334] "Generic (PLEG): container finished" podID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerID="2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278" exitCode=0 Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.116434 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" event={"ID":"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1","Type":"ContainerDied","Data":"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.116460 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" event={"ID":"5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1","Type":"ContainerDied","Data":"f56f8ee890a4b5247755ae1176188050f33cc05fb62d4484debd0a98d82d4503"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.116518 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-ctmrv" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.127064 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.127867 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9wbnx" event={"ID":"82778b9a-f721-4dce-ad71-9d36af2248ae","Type":"ContainerDied","Data":"5b19a6c44a5ac8256870edb2ad2d1a789425b36f9112a06f89352bcbb6d014ef"} Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.200682 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.201396 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-sth4d"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.215218 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.234593 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-ctmrv"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.253750 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.260633 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9wbnx"] Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.271876 4989 scope.go:117] "RemoveContainer" containerID="aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.312513 4989 scope.go:117] "RemoveContainer" containerID="bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962" Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.312924 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962\": container with ID starting with bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962 not found: ID does not exist" containerID="bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.312956 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962"} err="failed to get container status \"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962\": rpc error: code = NotFound desc = could not find container \"bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962\": container with ID starting with bf59a26c70b6aafe2c9b2c8f59840264f9a7a7a41e96bdf4bc1734cb0a3f6962 not found: ID does not exist" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.312983 4989 scope.go:117] "RemoveContainer" containerID="aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd" Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.313210 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd\": container with ID starting with aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd not found: ID does not exist" containerID="aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.313237 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd"} err="failed to get container status \"aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd\": rpc error: code = NotFound desc = could not find container \"aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd\": container with ID starting with aa3c24aa983dfe6b75e169e1b25469dea01e72fe014703ec5c253c3987e6eabd not found: ID does not exist" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.313255 4989 scope.go:117] "RemoveContainer" containerID="2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.466753 4989 scope.go:117] "RemoveContainer" containerID="36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.511342 4989 scope.go:117] "RemoveContainer" containerID="2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278" Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.511845 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278\": container with ID starting with 2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278 not found: ID does not exist" containerID="2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.511888 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278"} err="failed to get container status \"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278\": rpc error: code = NotFound desc = could not find container \"2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278\": container with ID starting with 2fb9e35de7ea152302b40d6507b307c27583f17788710cd4c6f393ddfd3c3278 not found: ID does not exist" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.511917 4989 scope.go:117] "RemoveContainer" containerID="36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5" Oct 06 09:00:07 crc kubenswrapper[4989]: E1006 09:00:07.512754 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5\": container with ID starting with 36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5 not found: ID does not exist" containerID="36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.512788 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5"} err="failed to get container status \"36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5\": rpc error: code = NotFound desc = could not find container \"36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5\": container with ID starting with 36fecaf277badccc9b892f8c3fd585c09711702a3df0f52bbbc8df3df479eea5 not found: ID does not exist" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.512802 4989 scope.go:117] "RemoveContainer" containerID="ffdb346adf2d6a16ff638180670a9c70883f99f9c7546b731a78608c9c33cfa6" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.946425 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" path="/var/lib/kubelet/pods/5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1/volumes" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.947155 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82778b9a-f721-4dce-ad71-9d36af2248ae" path="/var/lib/kubelet/pods/82778b9a-f721-4dce-ad71-9d36af2248ae/volumes" Oct 06 09:00:07 crc kubenswrapper[4989]: I1006 09:00:07.947742 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" path="/var/lib/kubelet/pods/db6cad76-cb75-4256-8a45-3eb27555e8a2/volumes" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.057954 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.058148 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.058167 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.058211 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:00:10.058196077 +0000 UTC m=+1260.848221657 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : configmap "swift-ring-files" not found Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.136708 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zzclq" event={"ID":"95f360dc-50e0-4b88-b25f-cc838c3a95f4","Type":"ContainerStarted","Data":"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26"} Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.136878 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.140003 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" event={"ID":"10a76137-a1aa-43ca-8c76-26aef58596ae","Type":"ContainerStarted","Data":"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251"} Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.140203 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.142619 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerStarted","Data":"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d"} Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.142664 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerStarted","Data":"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be"} Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.142765 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.157749 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-zzclq" podStartSLOduration=3.157731362 podStartE2EDuration="3.157731362s" podCreationTimestamp="2025-10-06 09:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:08.154088227 +0000 UTC m=+1258.944113847" watchObservedRunningTime="2025-10-06 09:00:08.157731362 +0000 UTC m=+1258.947756942" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.177765 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.245863969 podStartE2EDuration="4.177738807s" podCreationTimestamp="2025-10-06 09:00:04 +0000 UTC" firstStartedPulling="2025-10-06 09:00:05.406299888 +0000 UTC m=+1256.196325468" lastFinishedPulling="2025-10-06 09:00:07.338174726 +0000 UTC m=+1258.128200306" observedRunningTime="2025-10-06 09:00:08.171171258 +0000 UTC m=+1258.961196838" watchObservedRunningTime="2025-10-06 09:00:08.177738807 +0000 UTC m=+1258.967764387" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.191832 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6494d"] Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192311 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192335 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192361 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82778b9a-f721-4dce-ad71-9d36af2248ae" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192369 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="82778b9a-f721-4dce-ad71-9d36af2248ae" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192383 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192391 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192407 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192415 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192433 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192440 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: E1006 09:00:08.192469 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73589ab3-b75d-4d92-b312-2464ba72317a" containerName="mariadb-database-create" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192477 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="73589ab3-b75d-4d92-b312-2464ba72317a" containerName="mariadb-database-create" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192687 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b74bd43-f340-46e0-9c6d-db1b9ca7cbd1" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192705 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6cad76-cb75-4256-8a45-3eb27555e8a2" containerName="dnsmasq-dns" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192725 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="73589ab3-b75d-4d92-b312-2464ba72317a" containerName="mariadb-database-create" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.192740 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="82778b9a-f721-4dce-ad71-9d36af2248ae" containerName="init" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.193418 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6494d" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.196579 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" podStartSLOduration=4.196559289 podStartE2EDuration="4.196559289s" podCreationTimestamp="2025-10-06 09:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:08.188346633 +0000 UTC m=+1258.978372223" watchObservedRunningTime="2025-10-06 09:00:08.196559289 +0000 UTC m=+1258.986584869" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.203096 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6494d"] Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.261003 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnsdr\" (UniqueName: \"kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr\") pod \"glance-db-create-6494d\" (UID: \"ab4fe38d-017e-40fa-8d24-3da65f03839d\") " pod="openstack/glance-db-create-6494d" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.369438 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnsdr\" (UniqueName: \"kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr\") pod \"glance-db-create-6494d\" (UID: \"ab4fe38d-017e-40fa-8d24-3da65f03839d\") " pod="openstack/glance-db-create-6494d" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.409485 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnsdr\" (UniqueName: \"kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr\") pod \"glance-db-create-6494d\" (UID: \"ab4fe38d-017e-40fa-8d24-3da65f03839d\") " pod="openstack/glance-db-create-6494d" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.511824 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6494d" Oct 06 09:00:08 crc kubenswrapper[4989]: I1006 09:00:08.836507 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6494d"] Oct 06 09:00:08 crc kubenswrapper[4989]: W1006 09:00:08.847703 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab4fe38d_017e_40fa_8d24_3da65f03839d.slice/crio-08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c WatchSource:0}: Error finding container 08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c: Status 404 returned error can't find the container with id 08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c Oct 06 09:00:09 crc kubenswrapper[4989]: I1006 09:00:09.152130 4989 generic.go:334] "Generic (PLEG): container finished" podID="ab4fe38d-017e-40fa-8d24-3da65f03839d" containerID="0337cf022056d72d28b84790f3e2c0d6ff357bbd527ef1b5bb1231fc471cd03f" exitCode=0 Oct 06 09:00:09 crc kubenswrapper[4989]: I1006 09:00:09.153319 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6494d" event={"ID":"ab4fe38d-017e-40fa-8d24-3da65f03839d","Type":"ContainerDied","Data":"0337cf022056d72d28b84790f3e2c0d6ff357bbd527ef1b5bb1231fc471cd03f"} Oct 06 09:00:09 crc kubenswrapper[4989]: I1006 09:00:09.153349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6494d" event={"ID":"ab4fe38d-017e-40fa-8d24-3da65f03839d","Type":"ContainerStarted","Data":"08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c"} Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.101585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:10 crc kubenswrapper[4989]: E1006 09:00:10.101826 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:00:10 crc kubenswrapper[4989]: E1006 09:00:10.101839 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 09:00:10 crc kubenswrapper[4989]: E1006 09:00:10.101882 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:00:14.101866893 +0000 UTC m=+1264.891892473 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : configmap "swift-ring-files" not found Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.137494 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-qkp2p"] Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.138506 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.140476 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.142165 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.142458 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.152070 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qkp2p"] Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.307982 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308054 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4845\" (UniqueName: \"kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308108 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308156 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308194 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308249 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.308344 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409564 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409619 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4845\" (UniqueName: \"kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409669 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409696 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409749 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.409825 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.410364 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.410806 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.410923 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.414963 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.415672 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.416484 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.429078 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4845\" (UniqueName: \"kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845\") pod \"swift-ring-rebalance-qkp2p\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.464582 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-qszr8" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.473331 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.490997 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6494d" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.612833 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnsdr\" (UniqueName: \"kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr\") pod \"ab4fe38d-017e-40fa-8d24-3da65f03839d\" (UID: \"ab4fe38d-017e-40fa-8d24-3da65f03839d\") " Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.616887 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr" (OuterVolumeSpecName: "kube-api-access-jnsdr") pod "ab4fe38d-017e-40fa-8d24-3da65f03839d" (UID: "ab4fe38d-017e-40fa-8d24-3da65f03839d"). InnerVolumeSpecName "kube-api-access-jnsdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.715191 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnsdr\" (UniqueName: \"kubernetes.io/projected/ab4fe38d-017e-40fa-8d24-3da65f03839d-kube-api-access-jnsdr\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:10 crc kubenswrapper[4989]: I1006 09:00:10.928275 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qkp2p"] Oct 06 09:00:10 crc kubenswrapper[4989]: W1006 09:00:10.932012 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b32cd31_074d_4d27_aae2_dbcc01c6810b.slice/crio-07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc WatchSource:0}: Error finding container 07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc: Status 404 returned error can't find the container with id 07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc Oct 06 09:00:11 crc kubenswrapper[4989]: I1006 09:00:11.169050 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qkp2p" event={"ID":"3b32cd31-074d-4d27-aae2-dbcc01c6810b","Type":"ContainerStarted","Data":"07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc"} Oct 06 09:00:11 crc kubenswrapper[4989]: I1006 09:00:11.170924 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6494d" event={"ID":"ab4fe38d-017e-40fa-8d24-3da65f03839d","Type":"ContainerDied","Data":"08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c"} Oct 06 09:00:11 crc kubenswrapper[4989]: I1006 09:00:11.170956 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08423534d79189c18fc932b9cdd74fe8bfc4dbc493cfe4e02abbd84ae108461c" Oct 06 09:00:11 crc kubenswrapper[4989]: I1006 09:00:11.170970 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6494d" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.631711 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-551f-account-create-n8p8h"] Oct 06 09:00:12 crc kubenswrapper[4989]: E1006 09:00:12.632284 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4fe38d-017e-40fa-8d24-3da65f03839d" containerName="mariadb-database-create" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.632295 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4fe38d-017e-40fa-8d24-3da65f03839d" containerName="mariadb-database-create" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.632472 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4fe38d-017e-40fa-8d24-3da65f03839d" containerName="mariadb-database-create" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.633310 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.638174 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.639353 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-551f-account-create-n8p8h"] Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.750119 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pdxt\" (UniqueName: \"kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt\") pod \"keystone-551f-account-create-n8p8h\" (UID: \"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5\") " pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.851820 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pdxt\" (UniqueName: \"kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt\") pod \"keystone-551f-account-create-n8p8h\" (UID: \"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5\") " pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.880129 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pdxt\" (UniqueName: \"kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt\") pod \"keystone-551f-account-create-n8p8h\" (UID: \"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5\") " pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.957737 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.994997 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ca6e-account-create-c8xvj"] Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.996263 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:12 crc kubenswrapper[4989]: I1006 09:00:12.998754 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 06 09:00:13 crc kubenswrapper[4989]: I1006 09:00:13.001555 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ca6e-account-create-c8xvj"] Oct 06 09:00:13 crc kubenswrapper[4989]: I1006 09:00:13.157441 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv72h\" (UniqueName: \"kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h\") pod \"placement-ca6e-account-create-c8xvj\" (UID: \"2137ce1b-b046-4b3b-ab0a-3566df04dd46\") " pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:13 crc kubenswrapper[4989]: I1006 09:00:13.258971 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv72h\" (UniqueName: \"kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h\") pod \"placement-ca6e-account-create-c8xvj\" (UID: \"2137ce1b-b046-4b3b-ab0a-3566df04dd46\") " pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:13 crc kubenswrapper[4989]: I1006 09:00:13.277721 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv72h\" (UniqueName: \"kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h\") pod \"placement-ca6e-account-create-c8xvj\" (UID: \"2137ce1b-b046-4b3b-ab0a-3566df04dd46\") " pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:13 crc kubenswrapper[4989]: I1006 09:00:13.323212 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:14 crc kubenswrapper[4989]: I1006 09:00:14.174797 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:14 crc kubenswrapper[4989]: E1006 09:00:14.174977 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:00:14 crc kubenswrapper[4989]: E1006 09:00:14.175002 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 06 09:00:14 crc kubenswrapper[4989]: E1006 09:00:14.175054 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:00:22.175038335 +0000 UTC m=+1272.965063915 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : configmap "swift-ring-files" not found Oct 06 09:00:14 crc kubenswrapper[4989]: W1006 09:00:14.580564 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda6f7ed0_2bf5_4fe3_bb88_23991b325dd5.slice/crio-a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa WatchSource:0}: Error finding container a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa: Status 404 returned error can't find the container with id a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa Oct 06 09:00:14 crc kubenswrapper[4989]: W1006 09:00:14.581932 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2137ce1b_b046_4b3b_ab0a_3566df04dd46.slice/crio-89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d WatchSource:0}: Error finding container 89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d: Status 404 returned error can't find the container with id 89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d Oct 06 09:00:14 crc kubenswrapper[4989]: I1006 09:00:14.582576 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-551f-account-create-n8p8h"] Oct 06 09:00:14 crc kubenswrapper[4989]: I1006 09:00:14.590211 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ca6e-account-create-c8xvj"] Oct 06 09:00:14 crc kubenswrapper[4989]: I1006 09:00:14.856037 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.238815 4989 generic.go:334] "Generic (PLEG): container finished" podID="2137ce1b-b046-4b3b-ab0a-3566df04dd46" containerID="0722bd521228744da04f6113db075a5725d5c5a71bf59d2e590dba6347e4a201" exitCode=0 Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.238864 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca6e-account-create-c8xvj" event={"ID":"2137ce1b-b046-4b3b-ab0a-3566df04dd46","Type":"ContainerDied","Data":"0722bd521228744da04f6113db075a5725d5c5a71bf59d2e590dba6347e4a201"} Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.238900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca6e-account-create-c8xvj" event={"ID":"2137ce1b-b046-4b3b-ab0a-3566df04dd46","Type":"ContainerStarted","Data":"89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d"} Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.241083 4989 generic.go:334] "Generic (PLEG): container finished" podID="da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" containerID="a2eaeb715dab747daba65704d4937d60bc6b107c498217cba8e70206fbcef182" exitCode=0 Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.241149 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-551f-account-create-n8p8h" event={"ID":"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5","Type":"ContainerDied","Data":"a2eaeb715dab747daba65704d4937d60bc6b107c498217cba8e70206fbcef182"} Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.241184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-551f-account-create-n8p8h" event={"ID":"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5","Type":"ContainerStarted","Data":"a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa"} Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.242805 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qkp2p" event={"ID":"3b32cd31-074d-4d27-aae2-dbcc01c6810b","Type":"ContainerStarted","Data":"36be71fb07cf2fc2a215efb620b9b16bfbaddc2690904b596fe8d3c21c1cef65"} Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.320634 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-qkp2p" podStartSLOduration=2.131359129 podStartE2EDuration="5.320619703s" podCreationTimestamp="2025-10-06 09:00:10 +0000 UTC" firstStartedPulling="2025-10-06 09:00:10.934919557 +0000 UTC m=+1261.724945137" lastFinishedPulling="2025-10-06 09:00:14.124180111 +0000 UTC m=+1264.914205711" observedRunningTime="2025-10-06 09:00:15.318801461 +0000 UTC m=+1266.108827061" watchObservedRunningTime="2025-10-06 09:00:15.320619703 +0000 UTC m=+1266.110645283" Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.739634 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.800861 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:15 crc kubenswrapper[4989]: I1006 09:00:15.801068 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="dnsmasq-dns" containerID="cri-o://2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251" gracePeriod=10 Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.251969 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.253171 4989 generic.go:334] "Generic (PLEG): container finished" podID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerID="2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251" exitCode=0 Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.253606 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" event={"ID":"10a76137-a1aa-43ca-8c76-26aef58596ae","Type":"ContainerDied","Data":"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251"} Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.253794 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" event={"ID":"10a76137-a1aa-43ca-8c76-26aef58596ae","Type":"ContainerDied","Data":"71e15b83d4a258b09a135f2430b3266db27906a94fe112d7449ad8e0fe9e8e81"} Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.253874 4989 scope.go:117] "RemoveContainer" containerID="2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.350621 4989 scope.go:117] "RemoveContainer" containerID="839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.409359 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config\") pod \"10a76137-a1aa-43ca-8c76-26aef58596ae\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.409508 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7pc8\" (UniqueName: \"kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8\") pod \"10a76137-a1aa-43ca-8c76-26aef58596ae\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.409555 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb\") pod \"10a76137-a1aa-43ca-8c76-26aef58596ae\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.409581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb\") pod \"10a76137-a1aa-43ca-8c76-26aef58596ae\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.409688 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc\") pod \"10a76137-a1aa-43ca-8c76-26aef58596ae\" (UID: \"10a76137-a1aa-43ca-8c76-26aef58596ae\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.426596 4989 scope.go:117] "RemoveContainer" containerID="2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251" Oct 06 09:00:16 crc kubenswrapper[4989]: E1006 09:00:16.430615 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251\": container with ID starting with 2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251 not found: ID does not exist" containerID="2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.432050 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251"} err="failed to get container status \"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251\": rpc error: code = NotFound desc = could not find container \"2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251\": container with ID starting with 2c31a3d3480d5fa808e8623751a690d09b63e4c768809791f985fda1ae4b5251 not found: ID does not exist" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.432081 4989 scope.go:117] "RemoveContainer" containerID="839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9" Oct 06 09:00:16 crc kubenswrapper[4989]: E1006 09:00:16.434938 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9\": container with ID starting with 839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9 not found: ID does not exist" containerID="839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.434971 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9"} err="failed to get container status \"839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9\": rpc error: code = NotFound desc = could not find container \"839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9\": container with ID starting with 839218435a63cd4aa05403f3cf6bea1aa14264a50d4d5f4be5fa1fa0fdee00a9 not found: ID does not exist" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.454275 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8" (OuterVolumeSpecName: "kube-api-access-r7pc8") pod "10a76137-a1aa-43ca-8c76-26aef58596ae" (UID: "10a76137-a1aa-43ca-8c76-26aef58596ae"). InnerVolumeSpecName "kube-api-access-r7pc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.504659 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "10a76137-a1aa-43ca-8c76-26aef58596ae" (UID: "10a76137-a1aa-43ca-8c76-26aef58596ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.504834 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "10a76137-a1aa-43ca-8c76-26aef58596ae" (UID: "10a76137-a1aa-43ca-8c76-26aef58596ae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.512125 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7pc8\" (UniqueName: \"kubernetes.io/projected/10a76137-a1aa-43ca-8c76-26aef58596ae-kube-api-access-r7pc8\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.512148 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.512157 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.519640 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config" (OuterVolumeSpecName: "config") pod "10a76137-a1aa-43ca-8c76-26aef58596ae" (UID: "10a76137-a1aa-43ca-8c76-26aef58596ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.539168 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10a76137-a1aa-43ca-8c76-26aef58596ae" (UID: "10a76137-a1aa-43ca-8c76-26aef58596ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.615291 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.615337 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10a76137-a1aa-43ca-8c76-26aef58596ae-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.737878 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.784712 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.818820 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv72h\" (UniqueName: \"kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h\") pod \"2137ce1b-b046-4b3b-ab0a-3566df04dd46\" (UID: \"2137ce1b-b046-4b3b-ab0a-3566df04dd46\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.821874 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h" (OuterVolumeSpecName: "kube-api-access-hv72h") pod "2137ce1b-b046-4b3b-ab0a-3566df04dd46" (UID: "2137ce1b-b046-4b3b-ab0a-3566df04dd46"). InnerVolumeSpecName "kube-api-access-hv72h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.922621 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pdxt\" (UniqueName: \"kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt\") pod \"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5\" (UID: \"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5\") " Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.923199 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv72h\" (UniqueName: \"kubernetes.io/projected/2137ce1b-b046-4b3b-ab0a-3566df04dd46-kube-api-access-hv72h\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:16 crc kubenswrapper[4989]: I1006 09:00:16.925619 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt" (OuterVolumeSpecName: "kube-api-access-7pdxt") pod "da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" (UID: "da6f7ed0-2bf5-4fe3-bb88-23991b325dd5"). InnerVolumeSpecName "kube-api-access-7pdxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.025042 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pdxt\" (UniqueName: \"kubernetes.io/projected/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5-kube-api-access-7pdxt\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.262597 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ca6e-account-create-c8xvj" event={"ID":"2137ce1b-b046-4b3b-ab0a-3566df04dd46","Type":"ContainerDied","Data":"89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d"} Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.262654 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89db988919b4b00b33ca3ae9df3f8b69ed1a9554cd1e2a29d68bbda7b8742d4d" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.262731 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ca6e-account-create-c8xvj" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.273631 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-551f-account-create-n8p8h" event={"ID":"da6f7ed0-2bf5-4fe3-bb88-23991b325dd5","Type":"ContainerDied","Data":"a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa"} Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.273697 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1977b0b51185936112a410652777a73e6955eac6a84de3694721b7e1d3652fa" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.273751 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-551f-account-create-n8p8h" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.276770 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-qg47h" Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.323000 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.327654 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-qg47h"] Oct 06 09:00:17 crc kubenswrapper[4989]: I1006 09:00:17.948057 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" path="/var/lib/kubelet/pods/10a76137-a1aa-43ca-8c76-26aef58596ae/volumes" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.313829 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec63-account-create-2r8tq"] Oct 06 09:00:18 crc kubenswrapper[4989]: E1006 09:00:18.314146 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="init" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314161 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="init" Oct 06 09:00:18 crc kubenswrapper[4989]: E1006 09:00:18.314183 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="dnsmasq-dns" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314192 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="dnsmasq-dns" Oct 06 09:00:18 crc kubenswrapper[4989]: E1006 09:00:18.314219 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2137ce1b-b046-4b3b-ab0a-3566df04dd46" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314227 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2137ce1b-b046-4b3b-ab0a-3566df04dd46" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: E1006 09:00:18.314237 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314243 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314396 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314408 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a76137-a1aa-43ca-8c76-26aef58596ae" containerName="dnsmasq-dns" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.314421 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2137ce1b-b046-4b3b-ab0a-3566df04dd46" containerName="mariadb-account-create" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.315106 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.321922 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.322823 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec63-account-create-2r8tq"] Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.445547 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65d5r\" (UniqueName: \"kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r\") pod \"glance-ec63-account-create-2r8tq\" (UID: \"604e87ea-9072-4967-b6b8-c717378a4db7\") " pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.547155 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65d5r\" (UniqueName: \"kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r\") pod \"glance-ec63-account-create-2r8tq\" (UID: \"604e87ea-9072-4967-b6b8-c717378a4db7\") " pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.564278 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65d5r\" (UniqueName: \"kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r\") pod \"glance-ec63-account-create-2r8tq\" (UID: \"604e87ea-9072-4967-b6b8-c717378a4db7\") " pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:18 crc kubenswrapper[4989]: I1006 09:00:18.636909 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:19 crc kubenswrapper[4989]: I1006 09:00:19.087192 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec63-account-create-2r8tq"] Oct 06 09:00:19 crc kubenswrapper[4989]: I1006 09:00:19.291275 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec63-account-create-2r8tq" event={"ID":"604e87ea-9072-4967-b6b8-c717378a4db7","Type":"ContainerStarted","Data":"53051d7596f202845c2976ec0ec37eba12fdd22934bafc7e1539416dcc344c6a"} Oct 06 09:00:19 crc kubenswrapper[4989]: I1006 09:00:19.831287 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 06 09:00:20 crc kubenswrapper[4989]: I1006 09:00:20.303439 4989 generic.go:334] "Generic (PLEG): container finished" podID="604e87ea-9072-4967-b6b8-c717378a4db7" containerID="4bc2c7953a81e332c86eef320b2122bd2683241d8c8673b379fa2d669b4fcd77" exitCode=0 Oct 06 09:00:20 crc kubenswrapper[4989]: I1006 09:00:20.303783 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec63-account-create-2r8tq" event={"ID":"604e87ea-9072-4967-b6b8-c717378a4db7","Type":"ContainerDied","Data":"4bc2c7953a81e332c86eef320b2122bd2683241d8c8673b379fa2d669b4fcd77"} Oct 06 09:00:21 crc kubenswrapper[4989]: I1006 09:00:21.635639 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:21 crc kubenswrapper[4989]: I1006 09:00:21.714073 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65d5r\" (UniqueName: \"kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r\") pod \"604e87ea-9072-4967-b6b8-c717378a4db7\" (UID: \"604e87ea-9072-4967-b6b8-c717378a4db7\") " Oct 06 09:00:21 crc kubenswrapper[4989]: I1006 09:00:21.721167 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r" (OuterVolumeSpecName: "kube-api-access-65d5r") pod "604e87ea-9072-4967-b6b8-c717378a4db7" (UID: "604e87ea-9072-4967-b6b8-c717378a4db7"). InnerVolumeSpecName "kube-api-access-65d5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:21 crc kubenswrapper[4989]: I1006 09:00:21.816427 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65d5r\" (UniqueName: \"kubernetes.io/projected/604e87ea-9072-4967-b6b8-c717378a4db7-kube-api-access-65d5r\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.223603 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.229460 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"swift-storage-0\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " pod="openstack/swift-storage-0" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.320258 4989 generic.go:334] "Generic (PLEG): container finished" podID="3b32cd31-074d-4d27-aae2-dbcc01c6810b" containerID="36be71fb07cf2fc2a215efb620b9b16bfbaddc2690904b596fe8d3c21c1cef65" exitCode=0 Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.320329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qkp2p" event={"ID":"3b32cd31-074d-4d27-aae2-dbcc01c6810b","Type":"ContainerDied","Data":"36be71fb07cf2fc2a215efb620b9b16bfbaddc2690904b596fe8d3c21c1cef65"} Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.321957 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec63-account-create-2r8tq" event={"ID":"604e87ea-9072-4967-b6b8-c717378a4db7","Type":"ContainerDied","Data":"53051d7596f202845c2976ec0ec37eba12fdd22934bafc7e1539416dcc344c6a"} Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.321981 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53051d7596f202845c2976ec0ec37eba12fdd22934bafc7e1539416dcc344c6a" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.321990 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec63-account-create-2r8tq" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.401293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 09:00:22 crc kubenswrapper[4989]: I1006 09:00:22.904485 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:00:22 crc kubenswrapper[4989]: W1006 09:00:22.911436 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaf6911f_dbaf_432c_9103_71f58a087a12.slice/crio-7209a166318b3f15faf2edd793ce1ca9f607f07e72732eb9a45314021ae21935 WatchSource:0}: Error finding container 7209a166318b3f15faf2edd793ce1ca9f607f07e72732eb9a45314021ae21935: Status 404 returned error can't find the container with id 7209a166318b3f15faf2edd793ce1ca9f607f07e72732eb9a45314021ae21935 Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.332768 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"7209a166318b3f15faf2edd793ce1ca9f607f07e72732eb9a45314021ae21935"} Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.467493 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-v76fj"] Oct 06 09:00:23 crc kubenswrapper[4989]: E1006 09:00:23.467850 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604e87ea-9072-4967-b6b8-c717378a4db7" containerName="mariadb-account-create" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.467865 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="604e87ea-9072-4967-b6b8-c717378a4db7" containerName="mariadb-account-create" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.468022 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="604e87ea-9072-4967-b6b8-c717378a4db7" containerName="mariadb-account-create" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.468611 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.470531 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-89g6f" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.470798 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.491562 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-v76fj"] Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.545086 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgbj7\" (UniqueName: \"kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.545550 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.545584 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.545640 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.648204 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgbj7\" (UniqueName: \"kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.648342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.648368 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.648414 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.656215 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.656865 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.657388 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.665553 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgbj7\" (UniqueName: \"kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7\") pod \"glance-db-sync-v76fj\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.728053 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.793504 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.850797 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.850877 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.850924 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.851033 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.851083 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.851225 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.851249 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4845\" (UniqueName: \"kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845\") pod \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\" (UID: \"3b32cd31-074d-4d27-aae2-dbcc01c6810b\") " Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.851768 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.852085 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.855477 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845" (OuterVolumeSpecName: "kube-api-access-m4845") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "kube-api-access-m4845". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.858963 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.877974 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.878787 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.880596 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts" (OuterVolumeSpecName: "scripts") pod "3b32cd31-074d-4d27-aae2-dbcc01c6810b" (UID: "3b32cd31-074d-4d27-aae2-dbcc01c6810b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961319 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961359 4989 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961371 4989 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961382 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4845\" (UniqueName: \"kubernetes.io/projected/3b32cd31-074d-4d27-aae2-dbcc01c6810b-kube-api-access-m4845\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961392 4989 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3b32cd31-074d-4d27-aae2-dbcc01c6810b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961401 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b32cd31-074d-4d27-aae2-dbcc01c6810b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:23 crc kubenswrapper[4989]: I1006 09:00:23.961409 4989 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3b32cd31-074d-4d27-aae2-dbcc01c6810b-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.354871 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130"} Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.355271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c"} Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.357909 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qkp2p" event={"ID":"3b32cd31-074d-4d27-aae2-dbcc01c6810b","Type":"ContainerDied","Data":"07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc"} Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.357957 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b13b85991f9bd1517d7cbe22e007e618b6d9fe3010afe0bbc76ca013e908dc" Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.358026 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qkp2p" Oct 06 09:00:24 crc kubenswrapper[4989]: I1006 09:00:24.441098 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-v76fj"] Oct 06 09:00:24 crc kubenswrapper[4989]: W1006 09:00:24.447674 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02a2fb2b_b301_444d_9e4b_f8105f3f229e.slice/crio-950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d WatchSource:0}: Error finding container 950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d: Status 404 returned error can't find the container with id 950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.369284 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751"} Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.369720 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9"} Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.371430 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v76fj" event={"ID":"02a2fb2b-b301-444d-9e4b-f8105f3f229e","Type":"ContainerStarted","Data":"950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d"} Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.374673 4989 generic.go:334] "Generic (PLEG): container finished" podID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerID="081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4" exitCode=0 Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.374749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerDied","Data":"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4"} Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.377989 4989 generic.go:334] "Generic (PLEG): container finished" podID="03d954db-7dc6-4921-b260-1c189b9492c2" containerID="a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b" exitCode=0 Oct 06 09:00:25 crc kubenswrapper[4989]: I1006 09:00:25.378037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerDied","Data":"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b"} Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.387938 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerStarted","Data":"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca"} Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.388622 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.390989 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerStarted","Data":"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032"} Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.391638 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.412331 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.338319442 podStartE2EDuration="58.412314544s" podCreationTimestamp="2025-10-06 08:59:28 +0000 UTC" firstStartedPulling="2025-10-06 08:59:30.621775929 +0000 UTC m=+1221.411801509" lastFinishedPulling="2025-10-06 08:59:51.695771031 +0000 UTC m=+1242.485796611" observedRunningTime="2025-10-06 09:00:26.411648125 +0000 UTC m=+1277.201676265" watchObservedRunningTime="2025-10-06 09:00:26.412314544 +0000 UTC m=+1277.202340124" Oct 06 09:00:26 crc kubenswrapper[4989]: I1006 09:00:26.442835 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.296235773 podStartE2EDuration="59.442815472s" podCreationTimestamp="2025-10-06 08:59:27 +0000 UTC" firstStartedPulling="2025-10-06 08:59:29.546735411 +0000 UTC m=+1220.336760991" lastFinishedPulling="2025-10-06 08:59:51.69331512 +0000 UTC m=+1242.483340690" observedRunningTime="2025-10-06 09:00:26.43406371 +0000 UTC m=+1277.224089310" watchObservedRunningTime="2025-10-06 09:00:26.442815472 +0000 UTC m=+1277.232841052" Oct 06 09:00:27 crc kubenswrapper[4989]: I1006 09:00:27.406582 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75"} Oct 06 09:00:27 crc kubenswrapper[4989]: I1006 09:00:27.406927 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3"} Oct 06 09:00:27 crc kubenswrapper[4989]: I1006 09:00:27.406944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64"} Oct 06 09:00:28 crc kubenswrapper[4989]: I1006 09:00:28.429458 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.448788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.449096 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.449107 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.449118 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.449126 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d"} Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.852888 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.854004 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rzncg" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" probeResult="failure" output=< Oct 06 09:00:29 crc kubenswrapper[4989]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 06 09:00:29 crc kubenswrapper[4989]: > Oct 06 09:00:29 crc kubenswrapper[4989]: I1006 09:00:29.854887 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.089928 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rzncg-config-wp4jq"] Oct 06 09:00:30 crc kubenswrapper[4989]: E1006 09:00:30.090345 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b32cd31-074d-4d27-aae2-dbcc01c6810b" containerName="swift-ring-rebalance" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.090362 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b32cd31-074d-4d27-aae2-dbcc01c6810b" containerName="swift-ring-rebalance" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.090627 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b32cd31-074d-4d27-aae2-dbcc01c6810b" containerName="swift-ring-rebalance" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.091541 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.095011 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.105029 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rzncg-config-wp4jq"] Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163183 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163525 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163549 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gzv8\" (UniqueName: \"kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163692 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.163740 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.265162 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.265237 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.265262 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.265525 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.265948 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.266117 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.266154 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gzv8\" (UniqueName: \"kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.266220 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.266285 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.266370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.268554 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.289362 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gzv8\" (UniqueName: \"kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8\") pod \"ovn-controller-rzncg-config-wp4jq\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.447302 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:30 crc kubenswrapper[4989]: I1006 09:00:30.471249 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636"} Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.497442 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerStarted","Data":"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c"} Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.546691 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.989317649 podStartE2EDuration="27.546670255s" podCreationTimestamp="2025-10-06 09:00:05 +0000 UTC" firstStartedPulling="2025-10-06 09:00:22.913594544 +0000 UTC m=+1273.703620124" lastFinishedPulling="2025-10-06 09:00:28.47094715 +0000 UTC m=+1279.260972730" observedRunningTime="2025-10-06 09:00:32.542436424 +0000 UTC m=+1283.332462004" watchObservedRunningTime="2025-10-06 09:00:32.546670255 +0000 UTC m=+1283.336695835" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.825455 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.826961 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.836447 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.836849 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920332 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920374 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8gc2\" (UniqueName: \"kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920413 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920498 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920536 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:32.920551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022518 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022611 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022644 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8gc2\" (UniqueName: \"kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.022704 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.023537 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.023678 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.023999 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.024270 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.024863 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.046724 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8gc2\" (UniqueName: \"kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2\") pod \"dnsmasq-dns-77585f5f8c-d7pj8\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.149212 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.935157 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.935224 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.944940 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.945588 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:33.945643 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050" gracePeriod=600 Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:34.400195 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rzncg-config-wp4jq"] Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:34.528097 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050" exitCode=0 Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:34.528156 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050"} Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:34.528194 4989 scope.go:117] "RemoveContainer" containerID="71beef5e596f06c3940a32f33e3b9378c0f561f325e665575c463885fee5cb5f" Oct 06 09:00:34 crc kubenswrapper[4989]: I1006 09:00:34.819199 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rzncg" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" probeResult="failure" output=< Oct 06 09:00:34 crc kubenswrapper[4989]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 06 09:00:34 crc kubenswrapper[4989]: > Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.006943 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.288839 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-2n8b2"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.291015 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.304395 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2n8b2"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.375057 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.398196 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lmkzl"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.399246 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.444391 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lmkzl"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.472959 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5thz7\" (UniqueName: \"kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7\") pod \"cinder-db-create-2n8b2\" (UID: \"4d14187c-d5ec-4f67-b648-8d3d99e833c1\") " pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.575084 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdqfb\" (UniqueName: \"kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb\") pod \"barbican-db-create-lmkzl\" (UID: \"1da63f86-cdd5-4199-88a1-b617db5c046b\") " pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.575190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5thz7\" (UniqueName: \"kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7\") pod \"cinder-db-create-2n8b2\" (UID: \"4d14187c-d5ec-4f67-b648-8d3d99e833c1\") " pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.590283 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zs457"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.591833 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zs457" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.605247 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec"} Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.608381 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zs457"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.612311 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" event={"ID":"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d","Type":"ContainerStarted","Data":"940cc5388c47c8e3fe9203fe62a5451b6ec82adb9c32d2a42c43a57db2f296f7"} Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.619181 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5thz7\" (UniqueName: \"kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7\") pod \"cinder-db-create-2n8b2\" (UID: \"4d14187c-d5ec-4f67-b648-8d3d99e833c1\") " pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.620670 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg-config-wp4jq" event={"ID":"057b20c0-c11c-4e08-a473-3d8de64d0f74","Type":"ContainerStarted","Data":"3dc88fb3bd8ee12db2b77ecab178b7c37436701311d624fc0d00ff7ddc15a05d"} Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.620745 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg-config-wp4jq" event={"ID":"057b20c0-c11c-4e08-a473-3d8de64d0f74","Type":"ContainerStarted","Data":"1727a6144db0732d35ccd37501c464bad842219382d63f2e8f2bd588e0b52e97"} Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.632273 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.675011 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-nlwvf"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.676338 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr5xd\" (UniqueName: \"kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd\") pod \"neutron-db-create-zs457\" (UID: \"278f3372-6bb1-4ae2-871a-80a084a21c8d\") " pod="openstack/neutron-db-create-zs457" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.676414 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdqfb\" (UniqueName: \"kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb\") pod \"barbican-db-create-lmkzl\" (UID: \"1da63f86-cdd5-4199-88a1-b617db5c046b\") " pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.678074 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.683524 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.683982 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.684104 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.684264 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vq87b" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.695912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nlwvf"] Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.697595 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rzncg-config-wp4jq" podStartSLOduration=9.697568922 podStartE2EDuration="9.697568922s" podCreationTimestamp="2025-10-06 09:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:39.683473606 +0000 UTC m=+1290.473499186" watchObservedRunningTime="2025-10-06 09:00:39.697568922 +0000 UTC m=+1290.487594502" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.699044 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdqfb\" (UniqueName: \"kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb\") pod \"barbican-db-create-lmkzl\" (UID: \"1da63f86-cdd5-4199-88a1-b617db5c046b\") " pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.723320 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.778096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.778390 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr5xd\" (UniqueName: \"kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd\") pod \"neutron-db-create-zs457\" (UID: \"278f3372-6bb1-4ae2-871a-80a084a21c8d\") " pod="openstack/neutron-db-create-zs457" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.778430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgvfn\" (UniqueName: \"kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.778485 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.809802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr5xd\" (UniqueName: \"kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd\") pod \"neutron-db-create-zs457\" (UID: \"278f3372-6bb1-4ae2-871a-80a084a21c8d\") " pod="openstack/neutron-db-create-zs457" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.876709 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rzncg" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.885094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.885143 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgvfn\" (UniqueName: \"kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.885166 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.892498 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.895852 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.921482 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zs457" Oct 06 09:00:39 crc kubenswrapper[4989]: I1006 09:00:39.929809 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgvfn\" (UniqueName: \"kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn\") pod \"keystone-db-sync-nlwvf\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.018892 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.072794 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.226344 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2n8b2"] Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.417599 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lmkzl"] Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.628582 4989 generic.go:334] "Generic (PLEG): container finished" podID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerID="eda3adaa085fa6bd25f86914c0771cd240d33cb7530e8ad7c21b3823b9dc3430" exitCode=0 Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.628673 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" event={"ID":"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d","Type":"ContainerDied","Data":"eda3adaa085fa6bd25f86914c0771cd240d33cb7530e8ad7c21b3823b9dc3430"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.634306 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v76fj" event={"ID":"02a2fb2b-b301-444d-9e4b-f8105f3f229e","Type":"ContainerStarted","Data":"ef0a28225fec2703d402a5d120a0cb3520c4282520d26b5cba5f5b3fce0d8cb7"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.637203 4989 generic.go:334] "Generic (PLEG): container finished" podID="4d14187c-d5ec-4f67-b648-8d3d99e833c1" containerID="4700f2988f629fa1cc64cc9b1c2d7c4c3f12b989c2563fc3915d0197974d5911" exitCode=0 Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.637273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2n8b2" event={"ID":"4d14187c-d5ec-4f67-b648-8d3d99e833c1","Type":"ContainerDied","Data":"4700f2988f629fa1cc64cc9b1c2d7c4c3f12b989c2563fc3915d0197974d5911"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.637313 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2n8b2" event={"ID":"4d14187c-d5ec-4f67-b648-8d3d99e833c1","Type":"ContainerStarted","Data":"5e5aa6417aed8dc6da97e5e0698fd70c9b29cee8b28c181bb86c9aec2d100486"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.641209 4989 generic.go:334] "Generic (PLEG): container finished" podID="057b20c0-c11c-4e08-a473-3d8de64d0f74" containerID="3dc88fb3bd8ee12db2b77ecab178b7c37436701311d624fc0d00ff7ddc15a05d" exitCode=0 Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.641319 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg-config-wp4jq" event={"ID":"057b20c0-c11c-4e08-a473-3d8de64d0f74","Type":"ContainerDied","Data":"3dc88fb3bd8ee12db2b77ecab178b7c37436701311d624fc0d00ff7ddc15a05d"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.645623 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lmkzl" event={"ID":"1da63f86-cdd5-4199-88a1-b617db5c046b","Type":"ContainerStarted","Data":"4171f48766da159f34b60e071af620e6d01dc004f4bc666240c769ad99c40d7f"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.646074 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lmkzl" event={"ID":"1da63f86-cdd5-4199-88a1-b617db5c046b","Type":"ContainerStarted","Data":"370fabc7b65ac8001497f967e125ee0a8bdba465a66c7c718d06c2ba79ca4b6b"} Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.647788 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zs457"] Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.683806 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-lmkzl" podStartSLOduration=1.683784465 podStartE2EDuration="1.683784465s" podCreationTimestamp="2025-10-06 09:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:40.674352393 +0000 UTC m=+1291.464377973" watchObservedRunningTime="2025-10-06 09:00:40.683784465 +0000 UTC m=+1291.473810045" Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.744836 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-v76fj" podStartSLOduration=3.301066562 podStartE2EDuration="17.744820431s" podCreationTimestamp="2025-10-06 09:00:23 +0000 UTC" firstStartedPulling="2025-10-06 09:00:24.450374491 +0000 UTC m=+1275.240400071" lastFinishedPulling="2025-10-06 09:00:38.89412836 +0000 UTC m=+1289.684153940" observedRunningTime="2025-10-06 09:00:40.739790446 +0000 UTC m=+1291.529816026" watchObservedRunningTime="2025-10-06 09:00:40.744820431 +0000 UTC m=+1291.534846011" Oct 06 09:00:40 crc kubenswrapper[4989]: I1006 09:00:40.759017 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nlwvf"] Oct 06 09:00:40 crc kubenswrapper[4989]: W1006 09:00:40.761740 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff325bd0_b2a8_4af2_b48f_b8d9f5908474.slice/crio-4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9 WatchSource:0}: Error finding container 4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9: Status 404 returned error can't find the container with id 4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9 Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.654558 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nlwvf" event={"ID":"ff325bd0-b2a8-4af2-b48f-b8d9f5908474","Type":"ContainerStarted","Data":"4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9"} Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.656612 4989 generic.go:334] "Generic (PLEG): container finished" podID="1da63f86-cdd5-4199-88a1-b617db5c046b" containerID="4171f48766da159f34b60e071af620e6d01dc004f4bc666240c769ad99c40d7f" exitCode=0 Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.656681 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lmkzl" event={"ID":"1da63f86-cdd5-4199-88a1-b617db5c046b","Type":"ContainerDied","Data":"4171f48766da159f34b60e071af620e6d01dc004f4bc666240c769ad99c40d7f"} Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.658845 4989 generic.go:334] "Generic (PLEG): container finished" podID="278f3372-6bb1-4ae2-871a-80a084a21c8d" containerID="47d832e9c2490821149fdcd781bb935c878b27774b65fc445c59b2365810ab2f" exitCode=0 Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.658896 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zs457" event={"ID":"278f3372-6bb1-4ae2-871a-80a084a21c8d","Type":"ContainerDied","Data":"47d832e9c2490821149fdcd781bb935c878b27774b65fc445c59b2365810ab2f"} Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.658912 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zs457" event={"ID":"278f3372-6bb1-4ae2-871a-80a084a21c8d","Type":"ContainerStarted","Data":"e852077d8b809be96d56c9a102b1c5c9b907230f25fcdf042f9ee1b1a2032ae9"} Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.661179 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" event={"ID":"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d","Type":"ContainerStarted","Data":"917ea35f2f22ce6a48189a77c3f2a4f69c29fda3ced9e97d7c3722b7b9d7ab61"} Oct 06 09:00:41 crc kubenswrapper[4989]: I1006 09:00:41.695009 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" podStartSLOduration=9.694985716 podStartE2EDuration="9.694985716s" podCreationTimestamp="2025-10-06 09:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:41.691363772 +0000 UTC m=+1292.481389352" watchObservedRunningTime="2025-10-06 09:00:41.694985716 +0000 UTC m=+1292.485011296" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.115109 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.122693 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.234930 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235014 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235165 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gzv8\" (UniqueName: \"kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235203 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235236 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235295 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5thz7\" (UniqueName: \"kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7\") pod \"4d14187c-d5ec-4f67-b648-8d3d99e833c1\" (UID: \"4d14187c-d5ec-4f67-b648-8d3d99e833c1\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235341 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run\") pod \"057b20c0-c11c-4e08-a473-3d8de64d0f74\" (UID: \"057b20c0-c11c-4e08-a473-3d8de64d0f74\") " Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235570 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235620 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run" (OuterVolumeSpecName: "var-run") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235706 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235753 4989 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235771 4989 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.235800 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.236455 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts" (OuterVolumeSpecName: "scripts") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.241171 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8" (OuterVolumeSpecName: "kube-api-access-9gzv8") pod "057b20c0-c11c-4e08-a473-3d8de64d0f74" (UID: "057b20c0-c11c-4e08-a473-3d8de64d0f74"). InnerVolumeSpecName "kube-api-access-9gzv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.245911 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7" (OuterVolumeSpecName: "kube-api-access-5thz7") pod "4d14187c-d5ec-4f67-b648-8d3d99e833c1" (UID: "4d14187c-d5ec-4f67-b648-8d3d99e833c1"). InnerVolumeSpecName "kube-api-access-5thz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.337128 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gzv8\" (UniqueName: \"kubernetes.io/projected/057b20c0-c11c-4e08-a473-3d8de64d0f74-kube-api-access-9gzv8\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.337168 4989 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.337181 4989 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/057b20c0-c11c-4e08-a473-3d8de64d0f74-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.337192 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5thz7\" (UniqueName: \"kubernetes.io/projected/4d14187c-d5ec-4f67-b648-8d3d99e833c1-kube-api-access-5thz7\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.337209 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/057b20c0-c11c-4e08-a473-3d8de64d0f74-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.670919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2n8b2" event={"ID":"4d14187c-d5ec-4f67-b648-8d3d99e833c1","Type":"ContainerDied","Data":"5e5aa6417aed8dc6da97e5e0698fd70c9b29cee8b28c181bb86c9aec2d100486"} Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.670969 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e5aa6417aed8dc6da97e5e0698fd70c9b29cee8b28c181bb86c9aec2d100486" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.670933 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2n8b2" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.673902 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg-config-wp4jq" event={"ID":"057b20c0-c11c-4e08-a473-3d8de64d0f74","Type":"ContainerDied","Data":"1727a6144db0732d35ccd37501c464bad842219382d63f2e8f2bd588e0b52e97"} Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.673937 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1727a6144db0732d35ccd37501c464bad842219382d63f2e8f2bd588e0b52e97" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.674157 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg-config-wp4jq" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.674562 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.808885 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rzncg-config-wp4jq"] Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.814378 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rzncg-config-wp4jq"] Oct 06 09:00:42 crc kubenswrapper[4989]: I1006 09:00:42.945244 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zs457" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.047697 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr5xd\" (UniqueName: \"kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd\") pod \"278f3372-6bb1-4ae2-871a-80a084a21c8d\" (UID: \"278f3372-6bb1-4ae2-871a-80a084a21c8d\") " Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.051975 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd" (OuterVolumeSpecName: "kube-api-access-gr5xd") pod "278f3372-6bb1-4ae2-871a-80a084a21c8d" (UID: "278f3372-6bb1-4ae2-871a-80a084a21c8d"). InnerVolumeSpecName "kube-api-access-gr5xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.102150 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.150349 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr5xd\" (UniqueName: \"kubernetes.io/projected/278f3372-6bb1-4ae2-871a-80a084a21c8d-kube-api-access-gr5xd\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.251969 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdqfb\" (UniqueName: \"kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb\") pod \"1da63f86-cdd5-4199-88a1-b617db5c046b\" (UID: \"1da63f86-cdd5-4199-88a1-b617db5c046b\") " Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.258061 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb" (OuterVolumeSpecName: "kube-api-access-bdqfb") pod "1da63f86-cdd5-4199-88a1-b617db5c046b" (UID: "1da63f86-cdd5-4199-88a1-b617db5c046b"). InnerVolumeSpecName "kube-api-access-bdqfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.354627 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdqfb\" (UniqueName: \"kubernetes.io/projected/1da63f86-cdd5-4199-88a1-b617db5c046b-kube-api-access-bdqfb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.682877 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zs457" event={"ID":"278f3372-6bb1-4ae2-871a-80a084a21c8d","Type":"ContainerDied","Data":"e852077d8b809be96d56c9a102b1c5c9b907230f25fcdf042f9ee1b1a2032ae9"} Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.682931 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e852077d8b809be96d56c9a102b1c5c9b907230f25fcdf042f9ee1b1a2032ae9" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.682998 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zs457" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.690966 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lmkzl" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.691760 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lmkzl" event={"ID":"1da63f86-cdd5-4199-88a1-b617db5c046b","Type":"ContainerDied","Data":"370fabc7b65ac8001497f967e125ee0a8bdba465a66c7c718d06c2ba79ca4b6b"} Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.691793 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="370fabc7b65ac8001497f967e125ee0a8bdba465a66c7c718d06c2ba79ca4b6b" Oct 06 09:00:43 crc kubenswrapper[4989]: I1006 09:00:43.946754 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057b20c0-c11c-4e08-a473-3d8de64d0f74" path="/var/lib/kubelet/pods/057b20c0-c11c-4e08-a473-3d8de64d0f74/volumes" Oct 06 09:00:46 crc kubenswrapper[4989]: I1006 09:00:46.726795 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nlwvf" event={"ID":"ff325bd0-b2a8-4af2-b48f-b8d9f5908474","Type":"ContainerStarted","Data":"9fa32a8a48db3070a364396328866252d3262f83875879eea172b131eb0e8b9b"} Oct 06 09:00:46 crc kubenswrapper[4989]: I1006 09:00:46.745006 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-nlwvf" podStartSLOduration=2.500252052 podStartE2EDuration="7.744982171s" podCreationTimestamp="2025-10-06 09:00:39 +0000 UTC" firstStartedPulling="2025-10-06 09:00:40.76806242 +0000 UTC m=+1291.558088010" lastFinishedPulling="2025-10-06 09:00:46.012792539 +0000 UTC m=+1296.802818129" observedRunningTime="2025-10-06 09:00:46.740621996 +0000 UTC m=+1297.530647596" watchObservedRunningTime="2025-10-06 09:00:46.744982171 +0000 UTC m=+1297.535007791" Oct 06 09:00:47 crc kubenswrapper[4989]: I1006 09:00:47.746352 4989 generic.go:334] "Generic (PLEG): container finished" podID="02a2fb2b-b301-444d-9e4b-f8105f3f229e" containerID="ef0a28225fec2703d402a5d120a0cb3520c4282520d26b5cba5f5b3fce0d8cb7" exitCode=0 Oct 06 09:00:47 crc kubenswrapper[4989]: I1006 09:00:47.746444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v76fj" event={"ID":"02a2fb2b-b301-444d-9e4b-f8105f3f229e","Type":"ContainerDied","Data":"ef0a28225fec2703d402a5d120a0cb3520c4282520d26b5cba5f5b3fce0d8cb7"} Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.151929 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.204723 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.205049 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-zzclq" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="dnsmasq-dns" containerID="cri-o://3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26" gracePeriod=10 Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.705314 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.756193 4989 generic.go:334] "Generic (PLEG): container finished" podID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerID="3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26" exitCode=0 Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.756271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zzclq" event={"ID":"95f360dc-50e0-4b88-b25f-cc838c3a95f4","Type":"ContainerDied","Data":"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26"} Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.756333 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zzclq" event={"ID":"95f360dc-50e0-4b88-b25f-cc838c3a95f4","Type":"ContainerDied","Data":"26321d24117b92712d3bbf3353a55f35166412d8bfecb155455cb49ea24b5869"} Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.756357 4989 scope.go:117] "RemoveContainer" containerID="3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.756286 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zzclq" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.777526 4989 scope.go:117] "RemoveContainer" containerID="6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.798433 4989 scope.go:117] "RemoveContainer" containerID="3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26" Oct 06 09:00:48 crc kubenswrapper[4989]: E1006 09:00:48.798923 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26\": container with ID starting with 3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26 not found: ID does not exist" containerID="3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.798969 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26"} err="failed to get container status \"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26\": rpc error: code = NotFound desc = could not find container \"3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26\": container with ID starting with 3b15c65518833393cbdcc6a537e5efecf99fe8d8ef76f11c1b8b7fdf0953bb26 not found: ID does not exist" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.799016 4989 scope.go:117] "RemoveContainer" containerID="6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad" Oct 06 09:00:48 crc kubenswrapper[4989]: E1006 09:00:48.799246 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad\": container with ID starting with 6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad not found: ID does not exist" containerID="6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.799273 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad"} err="failed to get container status \"6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad\": rpc error: code = NotFound desc = could not find container \"6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad\": container with ID starting with 6bb5b30769e3bb335d599d32673895d7e74f289ed71bfd073e4e246a6daa77ad not found: ID does not exist" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.844452 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config\") pod \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.844492 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb\") pod \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.844689 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9nnb\" (UniqueName: \"kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb\") pod \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.844740 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb\") pod \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.844767 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc\") pod \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\" (UID: \"95f360dc-50e0-4b88-b25f-cc838c3a95f4\") " Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.850325 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb" (OuterVolumeSpecName: "kube-api-access-k9nnb") pod "95f360dc-50e0-4b88-b25f-cc838c3a95f4" (UID: "95f360dc-50e0-4b88-b25f-cc838c3a95f4"). InnerVolumeSpecName "kube-api-access-k9nnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.888552 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config" (OuterVolumeSpecName: "config") pod "95f360dc-50e0-4b88-b25f-cc838c3a95f4" (UID: "95f360dc-50e0-4b88-b25f-cc838c3a95f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.896762 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95f360dc-50e0-4b88-b25f-cc838c3a95f4" (UID: "95f360dc-50e0-4b88-b25f-cc838c3a95f4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.900467 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95f360dc-50e0-4b88-b25f-cc838c3a95f4" (UID: "95f360dc-50e0-4b88-b25f-cc838c3a95f4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.901731 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95f360dc-50e0-4b88-b25f-cc838c3a95f4" (UID: "95f360dc-50e0-4b88-b25f-cc838c3a95f4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.946316 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9nnb\" (UniqueName: \"kubernetes.io/projected/95f360dc-50e0-4b88-b25f-cc838c3a95f4-kube-api-access-k9nnb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.946351 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.946362 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.946370 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:48 crc kubenswrapper[4989]: I1006 09:00:48.946379 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95f360dc-50e0-4b88-b25f-cc838c3a95f4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.065140 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.103470 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.109297 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zzclq"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.153702 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data\") pod \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.153802 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle\") pod \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.153890 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgbj7\" (UniqueName: \"kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7\") pod \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.154342 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data\") pod \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\" (UID: \"02a2fb2b-b301-444d-9e4b-f8105f3f229e\") " Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.167119 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "02a2fb2b-b301-444d-9e4b-f8105f3f229e" (UID: "02a2fb2b-b301-444d-9e4b-f8105f3f229e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.177225 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7" (OuterVolumeSpecName: "kube-api-access-sgbj7") pod "02a2fb2b-b301-444d-9e4b-f8105f3f229e" (UID: "02a2fb2b-b301-444d-9e4b-f8105f3f229e"). InnerVolumeSpecName "kube-api-access-sgbj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.189343 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02a2fb2b-b301-444d-9e4b-f8105f3f229e" (UID: "02a2fb2b-b301-444d-9e4b-f8105f3f229e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.210342 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data" (OuterVolumeSpecName: "config-data") pod "02a2fb2b-b301-444d-9e4b-f8105f3f229e" (UID: "02a2fb2b-b301-444d-9e4b-f8105f3f229e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.257726 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.257764 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.257774 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgbj7\" (UniqueName: \"kubernetes.io/projected/02a2fb2b-b301-444d-9e4b-f8105f3f229e-kube-api-access-sgbj7\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.257786 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02a2fb2b-b301-444d-9e4b-f8105f3f229e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335346 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f526-account-create-t8hsg"] Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335765 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="init" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335787 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="init" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335802 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="dnsmasq-dns" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335811 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="dnsmasq-dns" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335839 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278f3372-6bb1-4ae2-871a-80a084a21c8d" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335847 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="278f3372-6bb1-4ae2-871a-80a084a21c8d" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335871 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a2fb2b-b301-444d-9e4b-f8105f3f229e" containerName="glance-db-sync" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335880 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a2fb2b-b301-444d-9e4b-f8105f3f229e" containerName="glance-db-sync" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335901 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057b20c0-c11c-4e08-a473-3d8de64d0f74" containerName="ovn-config" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335908 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="057b20c0-c11c-4e08-a473-3d8de64d0f74" containerName="ovn-config" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335923 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d14187c-d5ec-4f67-b648-8d3d99e833c1" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335931 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d14187c-d5ec-4f67-b648-8d3d99e833c1" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: E1006 09:00:49.335939 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da63f86-cdd5-4199-88a1-b617db5c046b" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.335947 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da63f86-cdd5-4199-88a1-b617db5c046b" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336147 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a2fb2b-b301-444d-9e4b-f8105f3f229e" containerName="glance-db-sync" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336173 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="278f3372-6bb1-4ae2-871a-80a084a21c8d" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336189 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="057b20c0-c11c-4e08-a473-3d8de64d0f74" containerName="ovn-config" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336200 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da63f86-cdd5-4199-88a1-b617db5c046b" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336212 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d14187c-d5ec-4f67-b648-8d3d99e833c1" containerName="mariadb-database-create" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336227 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" containerName="dnsmasq-dns" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.336902 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.339421 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.352852 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f526-account-create-t8hsg"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.464546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlljm\" (UniqueName: \"kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm\") pod \"cinder-f526-account-create-t8hsg\" (UID: \"04d66a34-26bc-4c9a-b0e6-d4bc809507f9\") " pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.524704 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8c94-account-create-zhkkc"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.525728 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.527936 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.533912 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8c94-account-create-zhkkc"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.566279 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlljm\" (UniqueName: \"kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm\") pod \"cinder-f526-account-create-t8hsg\" (UID: \"04d66a34-26bc-4c9a-b0e6-d4bc809507f9\") " pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.583182 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlljm\" (UniqueName: \"kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm\") pod \"cinder-f526-account-create-t8hsg\" (UID: \"04d66a34-26bc-4c9a-b0e6-d4bc809507f9\") " pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.667527 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w49kz\" (UniqueName: \"kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz\") pod \"barbican-8c94-account-create-zhkkc\" (UID: \"227e2f8d-debb-4226-a415-73b3e9bedcd4\") " pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.716444 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ec71-account-create-ms5qt"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.717468 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.719488 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.725169 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.725483 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ec71-account-create-ms5qt"] Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.770753 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w49kz\" (UniqueName: \"kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz\") pod \"barbican-8c94-account-create-zhkkc\" (UID: \"227e2f8d-debb-4226-a415-73b3e9bedcd4\") " pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.773540 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v76fj" event={"ID":"02a2fb2b-b301-444d-9e4b-f8105f3f229e","Type":"ContainerDied","Data":"950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d"} Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.773586 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="950e0612c32fed5a8a8f975ba62d56d9b86080b317b2975b304db2f941a0b15d" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.773677 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v76fj" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.783768 4989 generic.go:334] "Generic (PLEG): container finished" podID="ff325bd0-b2a8-4af2-b48f-b8d9f5908474" containerID="9fa32a8a48db3070a364396328866252d3262f83875879eea172b131eb0e8b9b" exitCode=0 Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.783843 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nlwvf" event={"ID":"ff325bd0-b2a8-4af2-b48f-b8d9f5908474","Type":"ContainerDied","Data":"9fa32a8a48db3070a364396328866252d3262f83875879eea172b131eb0e8b9b"} Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.806625 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w49kz\" (UniqueName: \"kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz\") pod \"barbican-8c94-account-create-zhkkc\" (UID: \"227e2f8d-debb-4226-a415-73b3e9bedcd4\") " pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.874537 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5krjf\" (UniqueName: \"kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf\") pod \"neutron-ec71-account-create-ms5qt\" (UID: \"20284027-d662-48a8-b61c-3b089247c1ea\") " pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.875048 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.958107 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f360dc-50e0-4b88-b25f-cc838c3a95f4" path="/var/lib/kubelet/pods/95f360dc-50e0-4b88-b25f-cc838c3a95f4/volumes" Oct 06 09:00:49 crc kubenswrapper[4989]: I1006 09:00:49.978708 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5krjf\" (UniqueName: \"kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf\") pod \"neutron-ec71-account-create-ms5qt\" (UID: \"20284027-d662-48a8-b61c-3b089247c1ea\") " pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.009849 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5krjf\" (UniqueName: \"kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf\") pod \"neutron-ec71-account-create-ms5qt\" (UID: \"20284027-d662-48a8-b61c-3b089247c1ea\") " pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.034854 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.184155 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.185454 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.212473 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.286714 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.286811 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkptb\" (UniqueName: \"kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.286884 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.287051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.287170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.287202 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.301226 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f526-account-create-t8hsg"] Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.388793 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.388839 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.388876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.388941 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkptb\" (UniqueName: \"kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.389052 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.389126 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.390219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.390586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.390928 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.391484 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.391590 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.415514 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkptb\" (UniqueName: \"kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb\") pod \"dnsmasq-dns-7ff5475cc9-k6wkk\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.430189 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8c94-account-create-zhkkc"] Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.518822 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.611076 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ec71-account-create-ms5qt"] Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.809269 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec71-account-create-ms5qt" event={"ID":"20284027-d662-48a8-b61c-3b089247c1ea","Type":"ContainerStarted","Data":"73a1d6a1544a1d3a0f92a2eb8d39ec0d74705060013a79d77caacf592dd5e439"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.809549 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec71-account-create-ms5qt" event={"ID":"20284027-d662-48a8-b61c-3b089247c1ea","Type":"ContainerStarted","Data":"ef6cfc544b1dc2cfc74c0bea9a9b34f233a13d822be3ac115a19e3f0ede3bc48"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.811669 4989 generic.go:334] "Generic (PLEG): container finished" podID="04d66a34-26bc-4c9a-b0e6-d4bc809507f9" containerID="c2f6d460e7859022f2d7bbbb38123c418fd428a488e6bb639d0fbd444a72b562" exitCode=0 Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.811742 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f526-account-create-t8hsg" event={"ID":"04d66a34-26bc-4c9a-b0e6-d4bc809507f9","Type":"ContainerDied","Data":"c2f6d460e7859022f2d7bbbb38123c418fd428a488e6bb639d0fbd444a72b562"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.811778 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f526-account-create-t8hsg" event={"ID":"04d66a34-26bc-4c9a-b0e6-d4bc809507f9","Type":"ContainerStarted","Data":"b54fd962810939ef4a9b957b94d680798d1c493d7da6a7b6141193b587fd9bcc"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.814106 4989 generic.go:334] "Generic (PLEG): container finished" podID="227e2f8d-debb-4226-a415-73b3e9bedcd4" containerID="3c033b104a71382421eb5e78709f0cd6d1f8852588b2f3002b7703097aa35b09" exitCode=0 Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.814140 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8c94-account-create-zhkkc" event={"ID":"227e2f8d-debb-4226-a415-73b3e9bedcd4","Type":"ContainerDied","Data":"3c033b104a71382421eb5e78709f0cd6d1f8852588b2f3002b7703097aa35b09"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.814189 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8c94-account-create-zhkkc" event={"ID":"227e2f8d-debb-4226-a415-73b3e9bedcd4","Type":"ContainerStarted","Data":"66744bf5d9c3b3a31f0a1282de009c06d9262347b39f19f4c92aa4657a1a6644"} Oct 06 09:00:50 crc kubenswrapper[4989]: I1006 09:00:50.835232 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ec71-account-create-ms5qt" podStartSLOduration=1.835205834 podStartE2EDuration="1.835205834s" podCreationTimestamp="2025-10-06 09:00:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:50.82707785 +0000 UTC m=+1301.617103440" watchObservedRunningTime="2025-10-06 09:00:50.835205834 +0000 UTC m=+1301.625231414" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.042360 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.108726 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.202761 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgvfn\" (UniqueName: \"kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn\") pod \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.203090 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle\") pod \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.203141 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data\") pod \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\" (UID: \"ff325bd0-b2a8-4af2-b48f-b8d9f5908474\") " Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.208485 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn" (OuterVolumeSpecName: "kube-api-access-lgvfn") pod "ff325bd0-b2a8-4af2-b48f-b8d9f5908474" (UID: "ff325bd0-b2a8-4af2-b48f-b8d9f5908474"). InnerVolumeSpecName "kube-api-access-lgvfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.233875 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff325bd0-b2a8-4af2-b48f-b8d9f5908474" (UID: "ff325bd0-b2a8-4af2-b48f-b8d9f5908474"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.252965 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data" (OuterVolumeSpecName: "config-data") pod "ff325bd0-b2a8-4af2-b48f-b8d9f5908474" (UID: "ff325bd0-b2a8-4af2-b48f-b8d9f5908474"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.304906 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgvfn\" (UniqueName: \"kubernetes.io/projected/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-kube-api-access-lgvfn\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.304955 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.304973 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff325bd0-b2a8-4af2-b48f-b8d9f5908474-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.822643 4989 generic.go:334] "Generic (PLEG): container finished" podID="866e6b71-7573-4a7d-86b8-9408b195f135" containerID="0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5" exitCode=0 Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.822725 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" event={"ID":"866e6b71-7573-4a7d-86b8-9408b195f135","Type":"ContainerDied","Data":"0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5"} Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.822754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" event={"ID":"866e6b71-7573-4a7d-86b8-9408b195f135","Type":"ContainerStarted","Data":"338ddeb7a0ba64a18b1a0129f2dd0b8f283f8b37cef036cf6766bfe76f6a290f"} Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.826207 4989 generic.go:334] "Generic (PLEG): container finished" podID="20284027-d662-48a8-b61c-3b089247c1ea" containerID="73a1d6a1544a1d3a0f92a2eb8d39ec0d74705060013a79d77caacf592dd5e439" exitCode=0 Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.826277 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec71-account-create-ms5qt" event={"ID":"20284027-d662-48a8-b61c-3b089247c1ea","Type":"ContainerDied","Data":"73a1d6a1544a1d3a0f92a2eb8d39ec0d74705060013a79d77caacf592dd5e439"} Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.827893 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nlwvf" event={"ID":"ff325bd0-b2a8-4af2-b48f-b8d9f5908474","Type":"ContainerDied","Data":"4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9"} Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.827940 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fcf8020c95f31c0da12f8b8727944b11cf79497c37ba3f21cc3914ce5eb77d9" Oct 06 09:00:51 crc kubenswrapper[4989]: I1006 09:00:51.828023 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nlwvf" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.101013 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.151140 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:52 crc kubenswrapper[4989]: E1006 09:00:52.151571 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff325bd0-b2a8-4af2-b48f-b8d9f5908474" containerName="keystone-db-sync" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.151583 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff325bd0-b2a8-4af2-b48f-b8d9f5908474" containerName="keystone-db-sync" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.151773 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff325bd0-b2a8-4af2-b48f-b8d9f5908474" containerName="keystone-db-sync" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.152704 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.203707 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6n68w"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.204849 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.209729 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.210299 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.213213 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.213511 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vq87b" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.219763 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.220190 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6n68w"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221146 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221216 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221251 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221269 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221291 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9krq\" (UniqueName: \"kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.221311 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.311612 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333411 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcz8w\" (UniqueName: \"kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333496 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333541 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333571 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333606 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333668 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333717 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333749 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333771 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333801 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9krq\" (UniqueName: \"kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333825 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.333856 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.334850 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.335197 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.335496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.336034 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.349748 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.381905 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9krq\" (UniqueName: \"kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq\") pod \"dnsmasq-dns-5c5cc7c5ff-tfmdg\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.438020 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlljm\" (UniqueName: \"kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm\") pod \"04d66a34-26bc-4c9a-b0e6-d4bc809507f9\" (UID: \"04d66a34-26bc-4c9a-b0e6-d4bc809507f9\") " Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.439420 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.439618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.439841 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.440506 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcz8w\" (UniqueName: \"kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.440595 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.440675 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.444239 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.456129 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm" (OuterVolumeSpecName: "kube-api-access-mlljm") pod "04d66a34-26bc-4c9a-b0e6-d4bc809507f9" (UID: "04d66a34-26bc-4c9a-b0e6-d4bc809507f9"). InnerVolumeSpecName "kube-api-access-mlljm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.468270 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.471402 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.487113 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.489054 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.490130 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.507233 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcz8w\" (UniqueName: \"kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w\") pod \"keystone-bootstrap-6n68w\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.561110 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.566787 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlljm\" (UniqueName: \"kubernetes.io/projected/04d66a34-26bc-4c9a-b0e6-d4bc809507f9-kube-api-access-mlljm\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.590967 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6tthk"] Oct 06 09:00:52 crc kubenswrapper[4989]: E1006 09:00:52.591309 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d66a34-26bc-4c9a-b0e6-d4bc809507f9" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.591337 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d66a34-26bc-4c9a-b0e6-d4bc809507f9" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.591501 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d66a34-26bc-4c9a-b0e6-d4bc809507f9" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.592045 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.610285 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.610522 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qkxpl" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.610809 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.618496 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.632493 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6tthk"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.654065 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.669199 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.669296 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.669347 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrhs2\" (UniqueName: \"kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.669388 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.669440 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.728300 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:00:52 crc kubenswrapper[4989]: E1006 09:00:52.728859 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="227e2f8d-debb-4226-a415-73b3e9bedcd4" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.728877 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="227e2f8d-debb-4226-a415-73b3e9bedcd4" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.729072 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="227e2f8d-debb-4226-a415-73b3e9bedcd4" containerName="mariadb-account-create" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.729878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.734780 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.772715 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w49kz\" (UniqueName: \"kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz\") pod \"227e2f8d-debb-4226-a415-73b3e9bedcd4\" (UID: \"227e2f8d-debb-4226-a415-73b3e9bedcd4\") " Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.773100 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrhs2\" (UniqueName: \"kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.773137 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.773182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.773221 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.773269 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.774524 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.787078 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz" (OuterVolumeSpecName: "kube-api-access-w49kz") pod "227e2f8d-debb-4226-a415-73b3e9bedcd4" (UID: "227e2f8d-debb-4226-a415-73b3e9bedcd4"). InnerVolumeSpecName "kube-api-access-w49kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.787662 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.788787 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.791103 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.806300 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrhs2\" (UniqueName: \"kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2\") pod \"placement-db-sync-6tthk\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.860200 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8c94-account-create-zhkkc" event={"ID":"227e2f8d-debb-4226-a415-73b3e9bedcd4","Type":"ContainerDied","Data":"66744bf5d9c3b3a31f0a1282de009c06d9262347b39f19f4c92aa4657a1a6644"} Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.860254 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66744bf5d9c3b3a31f0a1282de009c06d9262347b39f19f4c92aa4657a1a6644" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.860340 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8c94-account-create-zhkkc" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.867889 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" event={"ID":"866e6b71-7573-4a7d-86b8-9408b195f135","Type":"ContainerStarted","Data":"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba"} Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.868150 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="dnsmasq-dns" containerID="cri-o://7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba" gracePeriod=10 Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.868706 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.876977 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877038 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877099 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877140 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877236 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxvc5\" (UniqueName: \"kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877340 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.877466 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w49kz\" (UniqueName: \"kubernetes.io/projected/227e2f8d-debb-4226-a415-73b3e9bedcd4-kube-api-access-w49kz\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.882378 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f526-account-create-t8hsg" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.882601 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f526-account-create-t8hsg" event={"ID":"04d66a34-26bc-4c9a-b0e6-d4bc809507f9","Type":"ContainerDied","Data":"b54fd962810939ef4a9b957b94d680798d1c493d7da6a7b6141193b587fd9bcc"} Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.882644 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b54fd962810939ef4a9b957b94d680798d1c493d7da6a7b6141193b587fd9bcc" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.897200 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" podStartSLOduration=2.897185456 podStartE2EDuration="2.897185456s" podCreationTimestamp="2025-10-06 09:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:52.893914652 +0000 UTC m=+1303.683940252" watchObservedRunningTime="2025-10-06 09:00:52.897185456 +0000 UTC m=+1303.687211036" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.920937 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6tthk" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.981573 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxvc5\" (UniqueName: \"kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.981971 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.982252 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.982353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.982483 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.982571 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.983209 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.983471 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.983879 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.984238 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:52 crc kubenswrapper[4989]: I1006 09:00:52.988236 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.007341 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxvc5\" (UniqueName: \"kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5\") pod \"dnsmasq-dns-8b5c85b87-pk6g5\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.060001 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.261844 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.330608 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.332077 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.339432 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-89g6f" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.340413 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.340562 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.345314 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.348082 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.352267 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.352432 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.354697 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.367623 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.393866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.393914 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.394177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.394234 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.394256 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jpt\" (UniqueName: \"kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.394301 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.394359 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.451406 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6n68w"] Oct 06 09:00:53 crc kubenswrapper[4989]: W1006 09:00:53.455652 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod243c07e0_cb42_4c4e_b6b5_7aff5c4c3610.slice/crio-251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82 WatchSource:0}: Error finding container 251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82: Status 404 returned error can't find the container with id 251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82 Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504327 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504378 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504412 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504440 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504513 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxlhh\" (UniqueName: \"kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504560 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504602 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504661 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504711 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jpt\" (UniqueName: \"kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504735 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504759 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504840 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504883 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.504928 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.505414 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.505595 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.508587 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.509648 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.510321 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.511068 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.515790 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.520399 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.524995 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.531159 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.540842 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.574785 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jpt\" (UniqueName: \"kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt\") pod \"glance-default-external-api-0\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.600039 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6tthk"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606262 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606315 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606339 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606363 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606397 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606434 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606458 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606491 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606531 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlw29\" (UniqueName: \"kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606555 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606574 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxlhh\" (UniqueName: \"kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606606 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606642 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.606752 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.607114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.607916 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.620115 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.620803 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.622888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.630618 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxlhh\" (UniqueName: \"kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.641088 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.651578 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts\") pod \"ceilometer-0\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.662322 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.707864 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5krjf\" (UniqueName: \"kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf\") pod \"20284027-d662-48a8-b61c-3b089247c1ea\" (UID: \"20284027-d662-48a8-b61c-3b089247c1ea\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.708342 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.708394 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.710160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.710484 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.710672 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlw29\" (UniqueName: \"kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.710754 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.710838 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.713516 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.715279 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.716785 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.716958 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.723481 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.739194 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf" (OuterVolumeSpecName: "kube-api-access-5krjf") pod "20284027-d662-48a8-b61c-3b089247c1ea" (UID: "20284027-d662-48a8-b61c-3b089247c1ea"). InnerVolumeSpecName "kube-api-access-5krjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.740830 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.741300 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.754074 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlw29\" (UniqueName: \"kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.784945 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.811961 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812149 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812190 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812236 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812269 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812339 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkptb\" (UniqueName: \"kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb\") pod \"866e6b71-7573-4a7d-86b8-9408b195f135\" (UID: \"866e6b71-7573-4a7d-86b8-9408b195f135\") " Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.812755 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5krjf\" (UniqueName: \"kubernetes.io/projected/20284027-d662-48a8-b61c-3b089247c1ea-kube-api-access-5krjf\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.820642 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.833797 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb" (OuterVolumeSpecName: "kube-api-access-tkptb") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "kube-api-access-tkptb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:53 crc kubenswrapper[4989]: W1006 09:00:53.834050 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cdc2aaf_9021_4371_98e5_84b2ab8f96c8.slice/crio-bd9c2c8f221d35ab52840cf2e4f66bd35ee39ffbff9ff6af1f1592f656695b79 WatchSource:0}: Error finding container bd9c2c8f221d35ab52840cf2e4f66bd35ee39ffbff9ff6af1f1592f656695b79: Status 404 returned error can't find the container with id bd9c2c8f221d35ab52840cf2e4f66bd35ee39ffbff9ff6af1f1592f656695b79 Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.903156 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ec71-account-create-ms5qt" event={"ID":"20284027-d662-48a8-b61c-3b089247c1ea","Type":"ContainerDied","Data":"ef6cfc544b1dc2cfc74c0bea9a9b34f233a13d822be3ac115a19e3f0ede3bc48"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.907026 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef6cfc544b1dc2cfc74c0bea9a9b34f233a13d822be3ac115a19e3f0ede3bc48" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.904416 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ec71-account-create-ms5qt" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.912542 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" event={"ID":"f61ad8aa-260b-4809-90b3-3218e577fe23","Type":"ContainerStarted","Data":"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.912592 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" event={"ID":"f61ad8aa-260b-4809-90b3-3218e577fe23","Type":"ContainerStarted","Data":"a552bd2f3b3613a94a2976fb52f30b812d9e3665d4039c8ca57f60d814babb7c"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.914585 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkptb\" (UniqueName: \"kubernetes.io/projected/866e6b71-7573-4a7d-86b8-9408b195f135-kube-api-access-tkptb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.916453 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6tthk" event={"ID":"a144d611-7769-46ce-8707-3814fb0a7aa7","Type":"ContainerStarted","Data":"ec59d5b9eaf17ac3e4a477ecfdd1bba432fd31ab0b7b47c9fba855c04da2c2a2"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.919186 4989 generic.go:334] "Generic (PLEG): container finished" podID="866e6b71-7573-4a7d-86b8-9408b195f135" containerID="7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba" exitCode=0 Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.919246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" event={"ID":"866e6b71-7573-4a7d-86b8-9408b195f135","Type":"ContainerDied","Data":"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.919275 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" event={"ID":"866e6b71-7573-4a7d-86b8-9408b195f135","Type":"ContainerDied","Data":"338ddeb7a0ba64a18b1a0129f2dd0b8f283f8b37cef036cf6766bfe76f6a290f"} Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.919296 4989 scope.go:117] "RemoveContainer" containerID="7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.919439 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-k6wkk" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.952395 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.960916 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config" (OuterVolumeSpecName: "config") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:53 crc kubenswrapper[4989]: I1006 09:00:53.964238 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.007162 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" event={"ID":"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8","Type":"ContainerStarted","Data":"bd9c2c8f221d35ab52840cf2e4f66bd35ee39ffbff9ff6af1f1592f656695b79"} Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.007310 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6n68w" event={"ID":"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610","Type":"ContainerStarted","Data":"bec9e0cb8bf60ca2e27812afea65e265672c3c12101dea95cbd9385d1fbfa24e"} Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.007393 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6n68w" event={"ID":"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610","Type":"ContainerStarted","Data":"251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82"} Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.007577 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.014824 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.015482 4989 scope.go:117] "RemoveContainer" containerID="0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.017522 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.038124 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.038205 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.029497 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6n68w" podStartSLOduration=2.029477772 podStartE2EDuration="2.029477772s" podCreationTimestamp="2025-10-06 09:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:54.023961953 +0000 UTC m=+1304.813987533" watchObservedRunningTime="2025-10-06 09:00:54.029477772 +0000 UTC m=+1304.819503352" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.018121 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.058477 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "866e6b71-7573-4a7d-86b8-9408b195f135" (UID: "866e6b71-7573-4a7d-86b8-9408b195f135"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.092468 4989 scope.go:117] "RemoveContainer" containerID="7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba" Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.093044 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba\": container with ID starting with 7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba not found: ID does not exist" containerID="7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.093084 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba"} err="failed to get container status \"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba\": rpc error: code = NotFound desc = could not find container \"7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba\": container with ID starting with 7ff77bb4391c6c3ed113545421c5eca9cac34081f11e461209f365f33135e1ba not found: ID does not exist" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.093113 4989 scope.go:117] "RemoveContainer" containerID="0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5" Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.093888 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5\": container with ID starting with 0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5 not found: ID does not exist" containerID="0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.093935 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5"} err="failed to get container status \"0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5\": rpc error: code = NotFound desc = could not find container \"0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5\": container with ID starting with 0e348e32853e11e91b2a866a549ee91944a705aaf758b29948a2a9a8cfd9c1d5 not found: ID does not exist" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.141592 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.141619 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/866e6b71-7573-4a7d-86b8-9408b195f135-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.274911 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.287859 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-k6wkk"] Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.378205 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:00:54 crc kubenswrapper[4989]: W1006 09:00:54.381263 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod686780d8_ab3e_4337_a03d_ad0db9d5fea6.slice/crio-26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c WatchSource:0}: Error finding container 26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c: Status 404 returned error can't find the container with id 26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.601190 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jl67c"] Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.601770 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20284027-d662-48a8-b61c-3b089247c1ea" containerName="mariadb-account-create" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.601790 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="20284027-d662-48a8-b61c-3b089247c1ea" containerName="mariadb-account-create" Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.601805 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="dnsmasq-dns" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.601812 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="dnsmasq-dns" Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.601834 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="init" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.601840 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="init" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.602046 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" containerName="dnsmasq-dns" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.602066 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="20284027-d662-48a8-b61c-3b089247c1ea" containerName="mariadb-account-create" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.602599 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.608688 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rbwjv" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.608869 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.609051 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.609687 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.613754 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.630525 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jl67c"] Oct 06 09:00:54 crc kubenswrapper[4989]: W1006 09:00:54.641671 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00af6114_9242_4e7c_8ac3_416b1ff6baca.slice/crio-5b0cd071ddc1ae43bfdd6d6aafd87c61599a34e700c876461142b5d579a3ac99 WatchSource:0}: Error finding container 5b0cd071ddc1ae43bfdd6d6aafd87c61599a34e700c876461142b5d579a3ac99: Status 404 returned error can't find the container with id 5b0cd071ddc1ae43bfdd6d6aafd87c61599a34e700c876461142b5d579a3ac99 Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650159 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650235 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650316 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650378 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9krq\" (UniqueName: \"kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650468 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650579 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config\") pod \"f61ad8aa-260b-4809-90b3-3218e577fe23\" (UID: \"f61ad8aa-260b-4809-90b3-3218e577fe23\") " Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650876 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650916 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.650980 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq64b\" (UniqueName: \"kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.651022 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.651074 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.651107 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.666468 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq" (OuterVolumeSpecName: "kube-api-access-z9krq") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "kube-api-access-z9krq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.686621 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.706301 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.712614 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.721346 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config" (OuterVolumeSpecName: "config") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.725301 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f61ad8aa-260b-4809-90b3-3218e577fe23" (UID: "f61ad8aa-260b-4809-90b3-3218e577fe23"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.752747 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753171 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753240 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753276 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq64b\" (UniqueName: \"kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753399 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753473 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753499 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753513 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753526 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753537 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f61ad8aa-260b-4809-90b3-3218e577fe23-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.753549 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9krq\" (UniqueName: \"kubernetes.io/projected/f61ad8aa-260b-4809-90b3-3218e577fe23-kube-api-access-z9krq\") on node \"crc\" DevicePath \"\"" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.754057 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.763006 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.769581 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.770137 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.774413 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq64b\" (UniqueName: \"kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.774827 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts\") pod \"cinder-db-sync-jl67c\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.813352 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.933791 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-wtzcf"] Oct 06 09:00:54 crc kubenswrapper[4989]: E1006 09:00:54.934245 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f61ad8aa-260b-4809-90b3-3218e577fe23" containerName="init" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.934260 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f61ad8aa-260b-4809-90b3-3218e577fe23" containerName="init" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.934414 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f61ad8aa-260b-4809-90b3-3218e577fe23" containerName="init" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.935900 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.942351 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-28kjm" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.952919 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 09:00:54 crc kubenswrapper[4989]: I1006 09:00:54.974643 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wtzcf"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.033099 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jl67c" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.064423 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerStarted","Data":"26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.067468 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwf8n\" (UniqueName: \"kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.067588 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.067713 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.127607 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4mdrv"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.128648 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.134404 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vc6f7" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.134573 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.140042 4989 generic.go:334] "Generic (PLEG): container finished" podID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerID="da81eb1d67027417c132028f854ced3db667f08d0a3664d21d7c35ad4a15dd45" exitCode=0 Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.140122 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" event={"ID":"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8","Type":"ContainerDied","Data":"da81eb1d67027417c132028f854ced3db667f08d0a3664d21d7c35ad4a15dd45"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.141454 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.163929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerStarted","Data":"5b0cd071ddc1ae43bfdd6d6aafd87c61599a34e700c876461142b5d579a3ac99"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.168654 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.168706 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwf8n\" (UniqueName: \"kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.168775 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.172138 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4mdrv"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.174998 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.178915 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerStarted","Data":"126c956fa69324d001ad20cf5ff61f02c2f1d3b8d6591e0e04b64fef14ec6436"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.180412 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.193096 4989 generic.go:334] "Generic (PLEG): container finished" podID="f61ad8aa-260b-4809-90b3-3218e577fe23" containerID="62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be" exitCode=0 Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.193624 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwf8n\" (UniqueName: \"kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n\") pod \"barbican-db-sync-wtzcf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.193731 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.193981 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" event={"ID":"f61ad8aa-260b-4809-90b3-3218e577fe23","Type":"ContainerDied","Data":"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.194016 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg" event={"ID":"f61ad8aa-260b-4809-90b3-3218e577fe23","Type":"ContainerDied","Data":"a552bd2f3b3613a94a2976fb52f30b812d9e3665d4039c8ca57f60d814babb7c"} Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.194036 4989 scope.go:117] "RemoveContainer" containerID="62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.250031 4989 scope.go:117] "RemoveContainer" containerID="62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be" Oct 06 09:00:55 crc kubenswrapper[4989]: E1006 09:00:55.251933 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be\": container with ID starting with 62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be not found: ID does not exist" containerID="62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.251986 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be"} err="failed to get container status \"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be\": rpc error: code = NotFound desc = could not find container \"62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be\": container with ID starting with 62474c464f0d53d1b7eec02ff54fde1f6d8addf90c7361e7373f33f3d4e884be not found: ID does not exist" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.272177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.272241 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9j9j\" (UniqueName: \"kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.272314 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.318373 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.326640 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-tfmdg"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.337072 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.376391 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.376605 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.376692 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9j9j\" (UniqueName: \"kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.401126 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.401224 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.403024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9j9j\" (UniqueName: \"kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j\") pod \"neutron-db-sync-4mdrv\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.474536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.523810 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jl67c"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.661610 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.747124 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.811538 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.986889 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866e6b71-7573-4a7d-86b8-9408b195f135" path="/var/lib/kubelet/pods/866e6b71-7573-4a7d-86b8-9408b195f135/volumes" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.988222 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f61ad8aa-260b-4809-90b3-3218e577fe23" path="/var/lib/kubelet/pods/f61ad8aa-260b-4809-90b3-3218e577fe23/volumes" Oct 06 09:00:55 crc kubenswrapper[4989]: I1006 09:00:55.988788 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wtzcf"] Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.129500 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4mdrv"] Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.233501 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerStarted","Data":"293912f3fefab3ea52c04fc0cbc8a86445f49ef67f2bc2ec27278a34efebdaf8"} Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.239942 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jl67c" event={"ID":"eef21e7e-a1fe-4525-8bbe-5a7adac861cb","Type":"ContainerStarted","Data":"cbcbbeb8b5e893ba6f00b0069dba56a6c124ff055d757040209167cce4991a99"} Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.252765 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" event={"ID":"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8","Type":"ContainerStarted","Data":"341bec32385cc748dec180764963fa3e36a4420c1a18bc286a91c46965966f9e"} Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.253964 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.262183 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wtzcf" event={"ID":"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf","Type":"ContainerStarted","Data":"bd14ca5d397edcae5d659f93522abbbf6b0d8d3bacf60d0a1289c2af3f883828"} Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.273213 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerStarted","Data":"035f91d38dc192d523703fa81988c0eb5125a06f1b63c99461bf3c49883ad233"} Oct 06 09:00:56 crc kubenswrapper[4989]: I1006 09:00:56.280570 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4mdrv" event={"ID":"1bbd7317-2922-431a-bbf2-a22515821224","Type":"ContainerStarted","Data":"e847dc89c362086ee1a5d5b32c440e6dccc9e51c0e97aaa80e1e1c9ff153f558"} Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.309788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerStarted","Data":"c6d2d6d34ef679612f601313213c70302260925b0c73a08758af27e449b56af2"} Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.310162 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-log" containerID="cri-o://293912f3fefab3ea52c04fc0cbc8a86445f49ef67f2bc2ec27278a34efebdaf8" gracePeriod=30 Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.310458 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-httpd" containerID="cri-o://c6d2d6d34ef679612f601313213c70302260925b0c73a08758af27e449b56af2" gracePeriod=30 Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.342004 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerStarted","Data":"316d6cc011d092b3ddc376cbf4d0f73db645ad8effcfef4c3215df2b5159f29d"} Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.342125 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-httpd" containerID="cri-o://316d6cc011d092b3ddc376cbf4d0f73db645ad8effcfef4c3215df2b5159f29d" gracePeriod=30 Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.342390 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-log" containerID="cri-o://035f91d38dc192d523703fa81988c0eb5125a06f1b63c99461bf3c49883ad233" gracePeriod=30 Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.343148 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.343120607 podStartE2EDuration="5.343120607s" podCreationTimestamp="2025-10-06 09:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:57.337168356 +0000 UTC m=+1308.127193956" watchObservedRunningTime="2025-10-06 09:00:57.343120607 +0000 UTC m=+1308.133146187" Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.344886 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" podStartSLOduration=5.344876268 podStartE2EDuration="5.344876268s" podCreationTimestamp="2025-10-06 09:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:56.278302492 +0000 UTC m=+1307.068328082" watchObservedRunningTime="2025-10-06 09:00:57.344876268 +0000 UTC m=+1308.134901848" Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.355331 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4mdrv" event={"ID":"1bbd7317-2922-431a-bbf2-a22515821224","Type":"ContainerStarted","Data":"b27261eec2fed4866d7d50179573af96501d6417fcce76932c9276012042a4f0"} Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.390242 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.390215113 podStartE2EDuration="5.390215113s" podCreationTimestamp="2025-10-06 09:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:57.372808832 +0000 UTC m=+1308.162834432" watchObservedRunningTime="2025-10-06 09:00:57.390215113 +0000 UTC m=+1308.180240683" Oct 06 09:00:57 crc kubenswrapper[4989]: I1006 09:00:57.404925 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4mdrv" podStartSLOduration=2.404804402 podStartE2EDuration="2.404804402s" podCreationTimestamp="2025-10-06 09:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:00:57.395089313 +0000 UTC m=+1308.185114893" watchObservedRunningTime="2025-10-06 09:00:57.404804402 +0000 UTC m=+1308.194829982" Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.365561 4989 generic.go:334] "Generic (PLEG): container finished" podID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerID="316d6cc011d092b3ddc376cbf4d0f73db645ad8effcfef4c3215df2b5159f29d" exitCode=0 Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.366062 4989 generic.go:334] "Generic (PLEG): container finished" podID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerID="035f91d38dc192d523703fa81988c0eb5125a06f1b63c99461bf3c49883ad233" exitCode=143 Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.365645 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerDied","Data":"316d6cc011d092b3ddc376cbf4d0f73db645ad8effcfef4c3215df2b5159f29d"} Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.366160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerDied","Data":"035f91d38dc192d523703fa81988c0eb5125a06f1b63c99461bf3c49883ad233"} Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.371929 4989 generic.go:334] "Generic (PLEG): container finished" podID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerID="c6d2d6d34ef679612f601313213c70302260925b0c73a08758af27e449b56af2" exitCode=0 Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.371961 4989 generic.go:334] "Generic (PLEG): container finished" podID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerID="293912f3fefab3ea52c04fc0cbc8a86445f49ef67f2bc2ec27278a34efebdaf8" exitCode=143 Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.372152 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerDied","Data":"c6d2d6d34ef679612f601313213c70302260925b0c73a08758af27e449b56af2"} Oct 06 09:00:58 crc kubenswrapper[4989]: I1006 09:00:58.372211 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerDied","Data":"293912f3fefab3ea52c04fc0cbc8a86445f49ef67f2bc2ec27278a34efebdaf8"} Oct 06 09:00:59 crc kubenswrapper[4989]: I1006 09:00:59.385223 4989 generic.go:334] "Generic (PLEG): container finished" podID="243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" containerID="bec9e0cb8bf60ca2e27812afea65e265672c3c12101dea95cbd9385d1fbfa24e" exitCode=0 Oct 06 09:00:59 crc kubenswrapper[4989]: I1006 09:00:59.385279 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6n68w" event={"ID":"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610","Type":"ContainerDied","Data":"bec9e0cb8bf60ca2e27812afea65e265672c3c12101dea95cbd9385d1fbfa24e"} Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.400977 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9","Type":"ContainerDied","Data":"126c956fa69324d001ad20cf5ff61f02c2f1d3b8d6591e0e04b64fef14ec6436"} Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.401311 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="126c956fa69324d001ad20cf5ff61f02c2f1d3b8d6591e0e04b64fef14ec6436" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.404352 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"686780d8-ab3e-4337-a03d-ad0db9d5fea6","Type":"ContainerDied","Data":"26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c"} Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.404383 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26803054c8d69deca1529170ab5fcd9df79b86c630c132d7a1dd861d392da82c" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.447040 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.460718 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540324 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540612 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540784 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8jpt\" (UniqueName: \"kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540828 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540901 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.540954 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\" (UID: \"686780d8-ab3e-4337-a03d-ad0db9d5fea6\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.542213 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs" (OuterVolumeSpecName: "logs") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.542567 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.548509 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts" (OuterVolumeSpecName: "scripts") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.548550 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt" (OuterVolumeSpecName: "kube-api-access-n8jpt") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "kube-api-access-n8jpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.554811 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.610974 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data" (OuterVolumeSpecName: "config-data") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.629025 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "686780d8-ab3e-4337-a03d-ad0db9d5fea6" (UID: "686780d8-ab3e-4337-a03d-ad0db9d5fea6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.642634 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.642745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.642810 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.642893 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.643004 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlw29\" (UniqueName: \"kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.643033 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.643079 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs\") pod \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\" (UID: \"4c884b2d-6e10-4911-bcc6-acfbe6ef63a9\") " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.643615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.643737 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs" (OuterVolumeSpecName: "logs") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.650943 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29" (OuterVolumeSpecName: "kube-api-access-dlw29") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "kube-api-access-dlw29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652366 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652412 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlw29\" (UniqueName: \"kubernetes.io/projected/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-kube-api-access-dlw29\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652430 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652444 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652484 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652496 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652508 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652523 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/686780d8-ab3e-4337-a03d-ad0db9d5fea6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652536 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/686780d8-ab3e-4337-a03d-ad0db9d5fea6-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652548 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8jpt\" (UniqueName: \"kubernetes.io/projected/686780d8-ab3e-4337-a03d-ad0db9d5fea6-kube-api-access-n8jpt\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.652794 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts" (OuterVolumeSpecName: "scripts") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.656800 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.671607 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.685233 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.717775 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data" (OuterVolumeSpecName: "config-data") pod "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" (UID: "4c884b2d-6e10-4911-bcc6-acfbe6ef63a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.754481 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.754532 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.754568 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.754580 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.754592 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.784724 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 06 09:01:00 crc kubenswrapper[4989]: I1006 09:01:00.856160 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.413378 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.413429 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.471466 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.506195 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.524625 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: E1006 09:01:01.525033 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525058 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: E1006 09:01:01.525071 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525078 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: E1006 09:01:01.525086 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525092 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: E1006 09:01:01.525106 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525112 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525257 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525266 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525281 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-log" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.525293 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" containerName="glance-httpd" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.526175 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.531640 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.531816 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.531851 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.532407 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-89g6f" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.535787 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.543368 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.551988 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.562701 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.564149 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.567408 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.574851 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.574881 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677254 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677295 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677351 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677375 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677405 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677429 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-884pl\" (UniqueName: \"kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677448 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltvw8\" (UniqueName: \"kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677481 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677497 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677517 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677533 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677571 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.677602 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779444 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779501 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779531 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779563 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779591 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-884pl\" (UniqueName: \"kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779613 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779628 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltvw8\" (UniqueName: \"kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779646 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779706 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779731 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779752 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779772 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779792 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.779831 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.780281 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.780732 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.780744 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.781195 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.781593 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.781940 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.784813 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.786548 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.786546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.789911 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.796946 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.797512 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.799326 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.802246 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.803896 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltvw8\" (UniqueName: \"kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.813514 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.815776 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-884pl\" (UniqueName: \"kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.835402 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.857762 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.888595 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.948996 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c884b2d-6e10-4911-bcc6-acfbe6ef63a9" path="/var/lib/kubelet/pods/4c884b2d-6e10-4911-bcc6-acfbe6ef63a9/volumes" Oct 06 09:01:01 crc kubenswrapper[4989]: I1006 09:01:01.950113 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="686780d8-ab3e-4337-a03d-ad0db9d5fea6" path="/var/lib/kubelet/pods/686780d8-ab3e-4337-a03d-ad0db9d5fea6/volumes" Oct 06 09:01:03 crc kubenswrapper[4989]: I1006 09:01:03.062735 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:01:03 crc kubenswrapper[4989]: I1006 09:01:03.139949 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:01:03 crc kubenswrapper[4989]: I1006 09:01:03.140204 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="dnsmasq-dns" containerID="cri-o://917ea35f2f22ce6a48189a77c3f2a4f69c29fda3ced9e97d7c3722b7b9d7ab61" gracePeriod=10 Oct 06 09:01:03 crc kubenswrapper[4989]: I1006 09:01:03.478071 4989 generic.go:334] "Generic (PLEG): container finished" podID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerID="917ea35f2f22ce6a48189a77c3f2a4f69c29fda3ced9e97d7c3722b7b9d7ab61" exitCode=0 Oct 06 09:01:03 crc kubenswrapper[4989]: I1006 09:01:03.478424 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" event={"ID":"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d","Type":"ContainerDied","Data":"917ea35f2f22ce6a48189a77c3f2a4f69c29fda3ced9e97d7c3722b7b9d7ab61"} Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.695693 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.863150 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.863253 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.863287 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.863326 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcz8w\" (UniqueName: \"kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.863364 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.864599 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data\") pod \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\" (UID: \"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610\") " Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.869475 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts" (OuterVolumeSpecName: "scripts") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.870374 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.870499 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.871479 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w" (OuterVolumeSpecName: "kube-api-access-lcz8w") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "kube-api-access-lcz8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.889813 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.899635 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data" (OuterVolumeSpecName: "config-data") pod "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" (UID: "243c07e0-cb42-4c4e-b6b5-7aff5c4c3610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966388 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966422 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966431 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcz8w\" (UniqueName: \"kubernetes.io/projected/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-kube-api-access-lcz8w\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966440 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966448 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:05 crc kubenswrapper[4989]: I1006 09:01:05.966457 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.514775 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6n68w" event={"ID":"243c07e0-cb42-4c4e-b6b5-7aff5c4c3610","Type":"ContainerDied","Data":"251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82"} Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.514819 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="251a7ff338a14347c5a52f791e5a0cda6a3eae4d913399b570fe14ee254a9c82" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.514880 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6n68w" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.871071 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6n68w"] Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.881267 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6n68w"] Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.972442 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s6ttq"] Oct 06 09:01:06 crc kubenswrapper[4989]: E1006 09:01:06.972843 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" containerName="keystone-bootstrap" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.972857 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" containerName="keystone-bootstrap" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.973089 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" containerName="keystone-bootstrap" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.973608 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.977044 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.977303 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vq87b" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.977418 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.977577 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981718 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981757 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981780 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhdd4\" (UniqueName: \"kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981798 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981872 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.981954 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:06 crc kubenswrapper[4989]: I1006 09:01:06.984869 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s6ttq"] Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083345 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083403 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083452 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhdd4\" (UniqueName: \"kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.083523 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.090276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.091370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.092030 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.092250 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.092893 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.099704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhdd4\" (UniqueName: \"kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4\") pod \"keystone-bootstrap-s6ttq\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.295354 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:07 crc kubenswrapper[4989]: I1006 09:01:07.950778 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="243c07e0-cb42-4c4e-b6b5-7aff5c4c3610" path="/var/lib/kubelet/pods/243c07e0-cb42-4c4e-b6b5-7aff5c4c3610/volumes" Oct 06 09:01:12 crc kubenswrapper[4989]: E1006 09:01:12.749878 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 06 09:01:12 crc kubenswrapper[4989]: E1006 09:01:12.750336 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9bh686h57dh89h58h57ch5f4h644h674hb7h59dh5bdh58dh698h679h66ch57dh5f4h5cdh5cdh7h64bh5chfdh5b9hdbh5cfh598hf9h654hc6h65bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pxlhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(00af6114-9242-4e7c-8ac3-416b1ff6baca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.858546 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.987989 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.988086 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.988129 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.988214 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.988249 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:12 crc kubenswrapper[4989]: I1006 09:01:12.988279 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8gc2\" (UniqueName: \"kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2\") pod \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\" (UID: \"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d\") " Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.012830 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2" (OuterVolumeSpecName: "kube-api-access-s8gc2") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "kube-api-access-s8gc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.034056 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.038167 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.039166 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config" (OuterVolumeSpecName: "config") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.042564 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.054902 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" (UID: "a3a45a16-a53b-4bb3-b6e5-f9a0122d431d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090372 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090400 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090411 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090420 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090427 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.090435 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8gc2\" (UniqueName: \"kubernetes.io/projected/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d-kube-api-access-s8gc2\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.151166 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.581009 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" event={"ID":"a3a45a16-a53b-4bb3-b6e5-f9a0122d431d","Type":"ContainerDied","Data":"940cc5388c47c8e3fe9203fe62a5451b6ec82adb9c32d2a42c43a57db2f296f7"} Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.581075 4989 scope.go:117] "RemoveContainer" containerID="917ea35f2f22ce6a48189a77c3f2a4f69c29fda3ced9e97d7c3722b7b9d7ab61" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.581035 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-d7pj8" Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.613602 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.620886 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-d7pj8"] Oct 06 09:01:13 crc kubenswrapper[4989]: I1006 09:01:13.945922 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" path="/var/lib/kubelet/pods/a3a45a16-a53b-4bb3-b6e5-f9a0122d431d/volumes" Oct 06 09:01:14 crc kubenswrapper[4989]: E1006 09:01:14.910005 4989 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 06 09:01:14 crc kubenswrapper[4989]: E1006 09:01:14.910526 4989 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bq64b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jl67c_openstack(eef21e7e-a1fe-4525-8bbe-5a7adac861cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 06 09:01:14 crc kubenswrapper[4989]: E1006 09:01:14.911750 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jl67c" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" Oct 06 09:01:14 crc kubenswrapper[4989]: I1006 09:01:14.917372 4989 scope.go:117] "RemoveContainer" containerID="eda3adaa085fa6bd25f86914c0771cd240d33cb7530e8ad7c21b3823b9dc3430" Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.461240 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s6ttq"] Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.467561 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.598962 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6tthk" event={"ID":"a144d611-7769-46ce-8707-3814fb0a7aa7","Type":"ContainerStarted","Data":"9526030eb4478f8d3f9e0d4d119b074bfb7ed1ad9556aa7940abe108adaf3afc"} Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.602724 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wtzcf" event={"ID":"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf","Type":"ContainerStarted","Data":"94109ac5a382b61cf0a1a9382b747b9d85dd07d814f0ac516a8ddf57eba0b542"} Oct 06 09:01:15 crc kubenswrapper[4989]: E1006 09:01:15.614575 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-jl67c" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.622807 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6tthk" podStartSLOduration=4.482438487 podStartE2EDuration="23.622785652s" podCreationTimestamp="2025-10-06 09:00:52 +0000 UTC" firstStartedPulling="2025-10-06 09:00:53.625827546 +0000 UTC m=+1304.415853126" lastFinishedPulling="2025-10-06 09:01:12.766174711 +0000 UTC m=+1323.556200291" observedRunningTime="2025-10-06 09:01:15.622639478 +0000 UTC m=+1326.412665198" watchObservedRunningTime="2025-10-06 09:01:15.622785652 +0000 UTC m=+1326.412811232" Oct 06 09:01:15 crc kubenswrapper[4989]: I1006 09:01:15.645668 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-wtzcf" podStartSLOduration=2.778569309 podStartE2EDuration="21.64563031s" podCreationTimestamp="2025-10-06 09:00:54 +0000 UTC" firstStartedPulling="2025-10-06 09:00:55.984459035 +0000 UTC m=+1306.774484615" lastFinishedPulling="2025-10-06 09:01:14.851520036 +0000 UTC m=+1325.641545616" observedRunningTime="2025-10-06 09:01:15.636720113 +0000 UTC m=+1326.426745683" watchObservedRunningTime="2025-10-06 09:01:15.64563031 +0000 UTC m=+1326.435655890" Oct 06 09:01:15 crc kubenswrapper[4989]: W1006 09:01:15.666736 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod708a669c_612f_4261_947e_21b21effd0f0.slice/crio-035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65 WatchSource:0}: Error finding container 035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65: Status 404 returned error can't find the container with id 035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65 Oct 06 09:01:15 crc kubenswrapper[4989]: W1006 09:01:15.669107 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f433a70_1070_4c6c_ad1a_fa6bd5fb78fe.slice/crio-daf5b8c9bba19f7a8e55d18507fc3f32a633710d7290a1c92ce4a8d2511e7802 WatchSource:0}: Error finding container daf5b8c9bba19f7a8e55d18507fc3f32a633710d7290a1c92ce4a8d2511e7802: Status 404 returned error can't find the container with id daf5b8c9bba19f7a8e55d18507fc3f32a633710d7290a1c92ce4a8d2511e7802 Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.298727 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:01:16 crc kubenswrapper[4989]: W1006 09:01:16.309932 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8817e72b_0ee6_4c49_973d_4d5de2baf242.slice/crio-e87e111fe66777464f64e3cdb17073626edd1338f141742efb09025d841a4215 WatchSource:0}: Error finding container e87e111fe66777464f64e3cdb17073626edd1338f141742efb09025d841a4215: Status 404 returned error can't find the container with id e87e111fe66777464f64e3cdb17073626edd1338f141742efb09025d841a4215 Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.615122 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerStarted","Data":"e87e111fe66777464f64e3cdb17073626edd1338f141742efb09025d841a4215"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.617525 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerStarted","Data":"7d67e6989f2726f5fff92491c13e5646677be4b188c274a06e93bde0e31ed8f4"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.617562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerStarted","Data":"daf5b8c9bba19f7a8e55d18507fc3f32a633710d7290a1c92ce4a8d2511e7802"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.629296 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerStarted","Data":"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.632802 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s6ttq" event={"ID":"708a669c-612f-4261-947e-21b21effd0f0","Type":"ContainerStarted","Data":"03207da3ec8bb49a944e98ed52107432f3635b4287bebfcfc5db7845bce4ce5e"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.632883 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s6ttq" event={"ID":"708a669c-612f-4261-947e-21b21effd0f0","Type":"ContainerStarted","Data":"035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65"} Oct 06 09:01:16 crc kubenswrapper[4989]: I1006 09:01:16.654284 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s6ttq" podStartSLOduration=10.654265857 podStartE2EDuration="10.654265857s" podCreationTimestamp="2025-10-06 09:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:16.650337864 +0000 UTC m=+1327.440363454" watchObservedRunningTime="2025-10-06 09:01:16.654265857 +0000 UTC m=+1327.444291437" Oct 06 09:01:17 crc kubenswrapper[4989]: I1006 09:01:17.646622 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerStarted","Data":"6b63e8b67a9d19855d3c01aa43cd7cfb864abb3738572d14b99446cf360b4da7"} Oct 06 09:01:17 crc kubenswrapper[4989]: I1006 09:01:17.650416 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerStarted","Data":"374465b29904cc9b24031dc73ee4754d240a2e11bccbf776e34fa10c9f5ff9fc"} Oct 06 09:01:17 crc kubenswrapper[4989]: I1006 09:01:17.673876 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=16.673857881 podStartE2EDuration="16.673857881s" podCreationTimestamp="2025-10-06 09:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:17.672768419 +0000 UTC m=+1328.462794009" watchObservedRunningTime="2025-10-06 09:01:17.673857881 +0000 UTC m=+1328.463883461" Oct 06 09:01:18 crc kubenswrapper[4989]: I1006 09:01:18.664489 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerStarted","Data":"db767707a8ce13eb95542aa72cf874e036f3e2c48fc44f5cc7fd81804478b34b"} Oct 06 09:01:18 crc kubenswrapper[4989]: I1006 09:01:18.710573 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.710547586 podStartE2EDuration="17.710547586s" podCreationTimestamp="2025-10-06 09:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:18.699375365 +0000 UTC m=+1329.489400945" watchObservedRunningTime="2025-10-06 09:01:18.710547586 +0000 UTC m=+1329.500573176" Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.674661 4989 generic.go:334] "Generic (PLEG): container finished" podID="1bbd7317-2922-431a-bbf2-a22515821224" containerID="b27261eec2fed4866d7d50179573af96501d6417fcce76932c9276012042a4f0" exitCode=0 Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.674712 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4mdrv" event={"ID":"1bbd7317-2922-431a-bbf2-a22515821224","Type":"ContainerDied","Data":"b27261eec2fed4866d7d50179573af96501d6417fcce76932c9276012042a4f0"} Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.676279 4989 generic.go:334] "Generic (PLEG): container finished" podID="708a669c-612f-4261-947e-21b21effd0f0" containerID="03207da3ec8bb49a944e98ed52107432f3635b4287bebfcfc5db7845bce4ce5e" exitCode=0 Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.676317 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s6ttq" event={"ID":"708a669c-612f-4261-947e-21b21effd0f0","Type":"ContainerDied","Data":"03207da3ec8bb49a944e98ed52107432f3635b4287bebfcfc5db7845bce4ce5e"} Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.677748 4989 generic.go:334] "Generic (PLEG): container finished" podID="a144d611-7769-46ce-8707-3814fb0a7aa7" containerID="9526030eb4478f8d3f9e0d4d119b074bfb7ed1ad9556aa7940abe108adaf3afc" exitCode=0 Oct 06 09:01:19 crc kubenswrapper[4989]: I1006 09:01:19.678516 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6tthk" event={"ID":"a144d611-7769-46ce-8707-3814fb0a7aa7","Type":"ContainerDied","Data":"9526030eb4478f8d3f9e0d4d119b074bfb7ed1ad9556aa7940abe108adaf3afc"} Oct 06 09:01:20 crc kubenswrapper[4989]: I1006 09:01:20.705891 4989 generic.go:334] "Generic (PLEG): container finished" podID="0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" containerID="94109ac5a382b61cf0a1a9382b747b9d85dd07d814f0ac516a8ddf57eba0b542" exitCode=0 Oct 06 09:01:20 crc kubenswrapper[4989]: I1006 09:01:20.706462 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wtzcf" event={"ID":"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf","Type":"ContainerDied","Data":"94109ac5a382b61cf0a1a9382b747b9d85dd07d814f0ac516a8ddf57eba0b542"} Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.147978 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.158908 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.182168 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6tthk" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245232 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config\") pod \"1bbd7317-2922-431a-bbf2-a22515821224\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245354 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245433 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245451 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle\") pod \"1bbd7317-2922-431a-bbf2-a22515821224\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245474 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245543 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9j9j\" (UniqueName: \"kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j\") pod \"1bbd7317-2922-431a-bbf2-a22515821224\" (UID: \"1bbd7317-2922-431a-bbf2-a22515821224\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245604 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhdd4\" (UniqueName: \"kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.245642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys\") pod \"708a669c-612f-4261-947e-21b21effd0f0\" (UID: \"708a669c-612f-4261-947e-21b21effd0f0\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.251211 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.251470 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts" (OuterVolumeSpecName: "scripts") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.251933 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.252039 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j" (OuterVolumeSpecName: "kube-api-access-r9j9j") pod "1bbd7317-2922-431a-bbf2-a22515821224" (UID: "1bbd7317-2922-431a-bbf2-a22515821224"). InnerVolumeSpecName "kube-api-access-r9j9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.253558 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4" (OuterVolumeSpecName: "kube-api-access-hhdd4") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "kube-api-access-hhdd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.270193 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bbd7317-2922-431a-bbf2-a22515821224" (UID: "1bbd7317-2922-431a-bbf2-a22515821224"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.270582 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config" (OuterVolumeSpecName: "config") pod "1bbd7317-2922-431a-bbf2-a22515821224" (UID: "1bbd7317-2922-431a-bbf2-a22515821224"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.275519 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data" (OuterVolumeSpecName: "config-data") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.279219 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "708a669c-612f-4261-947e-21b21effd0f0" (UID: "708a669c-612f-4261-947e-21b21effd0f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.346533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrhs2\" (UniqueName: \"kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2\") pod \"a144d611-7769-46ce-8707-3814fb0a7aa7\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.346605 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data\") pod \"a144d611-7769-46ce-8707-3814fb0a7aa7\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.346646 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts\") pod \"a144d611-7769-46ce-8707-3814fb0a7aa7\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.346724 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs\") pod \"a144d611-7769-46ce-8707-3814fb0a7aa7\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.346759 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle\") pod \"a144d611-7769-46ce-8707-3814fb0a7aa7\" (UID: \"a144d611-7769-46ce-8707-3814fb0a7aa7\") " Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347284 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347308 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347321 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347336 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347347 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347358 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9j9j\" (UniqueName: \"kubernetes.io/projected/1bbd7317-2922-431a-bbf2-a22515821224-kube-api-access-r9j9j\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347370 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhdd4\" (UniqueName: \"kubernetes.io/projected/708a669c-612f-4261-947e-21b21effd0f0-kube-api-access-hhdd4\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347379 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/708a669c-612f-4261-947e-21b21effd0f0-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.347392 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bbd7317-2922-431a-bbf2-a22515821224-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.349414 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2" (OuterVolumeSpecName: "kube-api-access-hrhs2") pod "a144d611-7769-46ce-8707-3814fb0a7aa7" (UID: "a144d611-7769-46ce-8707-3814fb0a7aa7"). InnerVolumeSpecName "kube-api-access-hrhs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.349576 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs" (OuterVolumeSpecName: "logs") pod "a144d611-7769-46ce-8707-3814fb0a7aa7" (UID: "a144d611-7769-46ce-8707-3814fb0a7aa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.352011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts" (OuterVolumeSpecName: "scripts") pod "a144d611-7769-46ce-8707-3814fb0a7aa7" (UID: "a144d611-7769-46ce-8707-3814fb0a7aa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.372932 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data" (OuterVolumeSpecName: "config-data") pod "a144d611-7769-46ce-8707-3814fb0a7aa7" (UID: "a144d611-7769-46ce-8707-3814fb0a7aa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.375641 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a144d611-7769-46ce-8707-3814fb0a7aa7" (UID: "a144d611-7769-46ce-8707-3814fb0a7aa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.449389 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a144d611-7769-46ce-8707-3814fb0a7aa7-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.449440 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.449451 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrhs2\" (UniqueName: \"kubernetes.io/projected/a144d611-7769-46ce-8707-3814fb0a7aa7-kube-api-access-hrhs2\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.449461 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.449469 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a144d611-7769-46ce-8707-3814fb0a7aa7-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.730219 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4mdrv" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.731358 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4mdrv" event={"ID":"1bbd7317-2922-431a-bbf2-a22515821224","Type":"ContainerDied","Data":"e847dc89c362086ee1a5d5b32c440e6dccc9e51c0e97aaa80e1e1c9ff153f558"} Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.731400 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e847dc89c362086ee1a5d5b32c440e6dccc9e51c0e97aaa80e1e1c9ff153f558" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.736014 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s6ttq" event={"ID":"708a669c-612f-4261-947e-21b21effd0f0","Type":"ContainerDied","Data":"035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65"} Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.736062 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="035bc2d370896104086f07258a1a62bb58f71169498b6c07d57ed0330bfe6a65" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.736147 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s6ttq" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.742258 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6tthk" event={"ID":"a144d611-7769-46ce-8707-3814fb0a7aa7","Type":"ContainerDied","Data":"ec59d5b9eaf17ac3e4a477ecfdd1bba432fd31ab0b7b47c9fba855c04da2c2a2"} Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.742306 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec59d5b9eaf17ac3e4a477ecfdd1bba432fd31ab0b7b47c9fba855c04da2c2a2" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.742371 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6tthk" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.746918 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerStarted","Data":"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9"} Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.859628 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.861226 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.889964 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.890993 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.917737 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:01:21 crc kubenswrapper[4989]: E1006 09:01:21.918182 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bbd7317-2922-431a-bbf2-a22515821224" containerName="neutron-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918196 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bbd7317-2922-431a-bbf2-a22515821224" containerName="neutron-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: E1006 09:01:21.918216 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708a669c-612f-4261-947e-21b21effd0f0" containerName="keystone-bootstrap" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918224 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="708a669c-612f-4261-947e-21b21effd0f0" containerName="keystone-bootstrap" Oct 06 09:01:21 crc kubenswrapper[4989]: E1006 09:01:21.918241 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="init" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918249 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="init" Oct 06 09:01:21 crc kubenswrapper[4989]: E1006 09:01:21.918258 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a144d611-7769-46ce-8707-3814fb0a7aa7" containerName="placement-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918265 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a144d611-7769-46ce-8707-3814fb0a7aa7" containerName="placement-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: E1006 09:01:21.918280 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="dnsmasq-dns" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918287 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="dnsmasq-dns" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918479 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="708a669c-612f-4261-947e-21b21effd0f0" containerName="keystone-bootstrap" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918496 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a144d611-7769-46ce-8707-3814fb0a7aa7" containerName="placement-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918518 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a45a16-a53b-4bb3-b6e5-f9a0122d431d" containerName="dnsmasq-dns" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.918528 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bbd7317-2922-431a-bbf2-a22515821224" containerName="neutron-db-sync" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.919281 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.926284 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.926515 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.926668 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.926953 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.927219 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 09:01:21 crc kubenswrapper[4989]: I1006 09:01:21.927467 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vq87b" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.003750 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.003782 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005023 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005040 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005073 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005101 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005209 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.005253 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.006281 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.017292 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qkxpl" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.017696 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.019995 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.020258 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.020472 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.039391 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.059895 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.059975 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060033 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060071 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060150 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060193 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060279 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf95p\" (UniqueName: \"kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.060954 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061040 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xfbk\" (UniqueName: \"kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061093 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061153 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061181 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061262 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061311 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061407 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm27t\" (UniqueName: \"kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061432 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061486 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061601 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.061630 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.117091 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.121222 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.121372 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.129658 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.129898 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.130019 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vc6f7" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.130197 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.140980 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163425 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163477 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163519 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163545 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163575 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163608 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.163666 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164140 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164184 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf95p\" (UniqueName: \"kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164234 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164266 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xfbk\" (UniqueName: \"kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164285 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164308 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164330 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164348 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164371 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164391 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164427 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm27t\" (UniqueName: \"kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164443 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.164461 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.171560 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.171604 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.177260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.178814 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.179082 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.179088 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.185224 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.185242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.185861 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.186440 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.187615 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.189928 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.190049 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.190235 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.190571 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.192024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm27t\" (UniqueName: \"kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t\") pod \"dnsmasq-dns-84b966f6c9-4qt72\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.192574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.196172 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xfbk\" (UniqueName: \"kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.204423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf95p\" (UniqueName: \"kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.216412 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data\") pod \"keystone-6c7ff48d4c-577qm\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.221138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle\") pod \"placement-5d88db6f8d-sz2w2\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.265918 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.265964 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.265990 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crz8l\" (UniqueName: \"kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.266044 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.266108 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.290397 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.319214 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.343230 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.372534 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwf8n\" (UniqueName: \"kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n\") pod \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.372806 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle\") pod \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.376850 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data\") pod \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\" (UID: \"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf\") " Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.377497 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.377701 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.377734 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.377778 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crz8l\" (UniqueName: \"kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.377926 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.386858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" (UID: "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.387718 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.388105 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n" (OuterVolumeSpecName: "kube-api-access-rwf8n") pod "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" (UID: "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf"). InnerVolumeSpecName "kube-api-access-rwf8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.389934 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.390325 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.406369 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.406975 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.409643 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crz8l\" (UniqueName: \"kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l\") pod \"neutron-5c8d7c9f64-s9jkz\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.441444 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" (UID: "0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.445082 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.480452 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.480489 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwf8n\" (UniqueName: \"kubernetes.io/projected/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-kube-api-access-rwf8n\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.480500 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.762284 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wtzcf" event={"ID":"0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf","Type":"ContainerDied","Data":"bd14ca5d397edcae5d659f93522abbbf6b0d8d3bacf60d0a1289c2af3f883828"} Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.762338 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd14ca5d397edcae5d659f93522abbbf6b0d8d3bacf60d0a1289c2af3f883828" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.762407 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wtzcf" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.763483 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.763507 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.763519 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.763564 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.835893 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.987290 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:01:22 crc kubenswrapper[4989]: E1006 09:01:22.987793 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" containerName="barbican-db-sync" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.987818 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" containerName="barbican-db-sync" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.988035 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" containerName="barbican-db-sync" Oct 06 09:01:22 crc kubenswrapper[4989]: I1006 09:01:22.989255 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.000190 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.000427 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-28kjm" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.004419 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.014790 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.023713 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.025396 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.031333 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.036487 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.091745 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094250 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094297 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bb5v\" (UniqueName: \"kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094342 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094367 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094386 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094405 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m44gr\" (UniqueName: \"kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094421 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094461 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094483 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.094516 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.176661 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200344 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200396 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bb5v\" (UniqueName: \"kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200466 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200581 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200628 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200683 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m44gr\" (UniqueName: \"kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200717 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200780 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200811 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.200886 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.210153 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.210567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.213747 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.219734 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.220827 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.221598 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.223188 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.223196 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.227853 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.227942 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.229035 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bb5v\" (UniqueName: \"kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.230147 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom\") pod \"barbican-keystone-listener-84fbfd87f8-2zrs8\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.236412 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m44gr\" (UniqueName: \"kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr\") pod \"barbican-worker-7c9c89c9ff-5cnfn\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.242172 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.251139 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.253192 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.254994 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.264322 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.305451 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.305953 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.306173 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.306490 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.306634 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.306889 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdd26\" (UniqueName: \"kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.316782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.317006 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jrgc\" (UniqueName: \"kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.317248 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.317469 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.317687 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.347153 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.376506 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.391406 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419231 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419282 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419320 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419340 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419373 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419400 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdd26\" (UniqueName: \"kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419429 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419454 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jrgc\" (UniqueName: \"kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419547 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.419571 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.420876 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.421566 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.422647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.424341 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.425083 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.425358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.436780 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.442788 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jrgc\" (UniqueName: \"kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc\") pod \"dnsmasq-dns-75c8ddd69c-dcdln\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.443608 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdd26\" (UniqueName: \"kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.449324 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.449427 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle\") pod \"barbican-api-6db7d8b6f4-2xmnl\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.572085 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.625321 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.815786 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" event={"ID":"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0","Type":"ContainerStarted","Data":"469b8fe99e7f4875e9f39ec7b673807ae46c00f69cf0f900e47f5d4549d4e26d"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.817237 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerStarted","Data":"30f7b6ec4c5170c507992f7f853a08947aca784f43b83a2255a4f575a7248bec"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.824087 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerStarted","Data":"35ffcc7392dc41bc132668ec57014e8ac2e8039a4359f8333605eabdde751c1e"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.824142 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerStarted","Data":"3425ce4ec6dcc5718601388a7bdbf9d7dac148a51a4bf23672929d6c5676dbbb"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.852196 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7ff48d4c-577qm" event={"ID":"920b5c92-bd05-43f9-aae2-820bb7142635","Type":"ContainerStarted","Data":"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.852234 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7ff48d4c-577qm" event={"ID":"920b5c92-bd05-43f9-aae2-820bb7142635","Type":"ContainerStarted","Data":"2028c691029687175d5e45716e016814d0107edae5d8330358f6401cd71e286a"} Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.852656 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:23 crc kubenswrapper[4989]: I1006 09:01:23.890343 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c7ff48d4c-577qm" podStartSLOduration=2.890321015 podStartE2EDuration="2.890321015s" podCreationTimestamp="2025-10-06 09:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:23.884882848 +0000 UTC m=+1334.674908428" watchObservedRunningTime="2025-10-06 09:01:23.890321015 +0000 UTC m=+1334.680346595" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.153950 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.475186 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.666720 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.679265 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.867238 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerStarted","Data":"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.867609 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerStarted","Data":"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.870671 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerStarted","Data":"d0976df43c3e6df06b0524405ab2b91644a303bb21801f327628624b1e011733"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.882271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" event={"ID":"742c9029-45a8-47ad-9fac-cd2f6c76fc15","Type":"ContainerStarted","Data":"6d1e677e66cae4eb8904531e607c05735d35f720e24158776bda9497370bca65"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.884971 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerStarted","Data":"4a55740396b636c8de2207c1b20cbc61bd80b8803c4aaa8ddc33405ea2cecaed"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.885789 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.885839 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.905129 4989 generic.go:334] "Generic (PLEG): container finished" podID="44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" containerID="a71f31e76a248fbab03a20daf73589fa0f807eb114eb0d617872371ec91890f1" exitCode=0 Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.905195 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" event={"ID":"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0","Type":"ContainerDied","Data":"a71f31e76a248fbab03a20daf73589fa0f807eb114eb0d617872371ec91890f1"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.927854 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerStarted","Data":"5fd13a48c0ac4ff5bca90fd8daf5fdb73f10b10fcc260ff67cf9e738467accd7"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.930683 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5d88db6f8d-sz2w2" podStartSLOduration=3.930642595 podStartE2EDuration="3.930642595s" podCreationTimestamp="2025-10-06 09:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:24.91308931 +0000 UTC m=+1335.703114890" watchObservedRunningTime="2025-10-06 09:01:24.930642595 +0000 UTC m=+1335.720668175" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.931142 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.932616 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerStarted","Data":"8c2e64b1115e348874a19be0c8f6bb19bc04442309cc5a7bd10dc4c828b3c765"} Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.932747 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.932762 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.963478 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.970036 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.973198 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.973965 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 06 09:01:24 crc kubenswrapper[4989]: I1006 09:01:24.974102 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.089644 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090052 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090097 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090117 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090201 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhcwg\" (UniqueName: \"kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.090235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.195655 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhcwg\" (UniqueName: \"kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.195760 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.195909 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.195948 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.195999 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.196024 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.196081 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.220712 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.221192 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.222819 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhcwg\" (UniqueName: \"kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.223930 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.225098 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.226650 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.227905 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs\") pod \"neutron-557b4f9dd9-blscg\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.305219 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.313507 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.401275 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.402189 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.402235 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.402253 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.402271 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm27t\" (UniqueName: \"kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.402409 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc\") pod \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\" (UID: \"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0\") " Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.406947 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t" (OuterVolumeSpecName: "kube-api-access-hm27t") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "kube-api-access-hm27t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.434765 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.440011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.451856 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.455105 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.515783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config" (OuterVolumeSpecName: "config") pod "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" (UID: "44b676b3-eb01-45e2-a91f-cc9b57e3e2e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518304 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518335 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518349 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518360 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518374 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.518387 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm27t\" (UniqueName: \"kubernetes.io/projected/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0-kube-api-access-hm27t\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:25 crc kubenswrapper[4989]: I1006 09:01:25.956738 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.957845 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-4qt72" event={"ID":"44b676b3-eb01-45e2-a91f-cc9b57e3e2e0","Type":"ContainerDied","Data":"469b8fe99e7f4875e9f39ec7b673807ae46c00f69cf0f900e47f5d4549d4e26d"} Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.957941 4989 scope.go:117] "RemoveContainer" containerID="a71f31e76a248fbab03a20daf73589fa0f807eb114eb0d617872371ec91890f1" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.962527 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerStarted","Data":"e04948b0c9de5595097016939d537f442c08815515ea97b9b0edddca18a4e910"} Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.964852 4989 generic.go:334] "Generic (PLEG): container finished" podID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerID="25d25936563097e7fdc38f6b1fcc6b1675c765f066a2f410c419974bfe7533a9" exitCode=0 Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.964931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" event={"ID":"742c9029-45a8-47ad-9fac-cd2f6c76fc15","Type":"ContainerDied","Data":"25d25936563097e7fdc38f6b1fcc6b1675c765f066a2f410c419974bfe7533a9"} Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.965044 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:25.968670 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.012804 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c8d7c9f64-s9jkz" podStartSLOduration=4.012780748 podStartE2EDuration="4.012780748s" podCreationTimestamp="2025-10-06 09:01:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:26.00938676 +0000 UTC m=+1336.799412370" watchObservedRunningTime="2025-10-06 09:01:26.012780748 +0000 UTC m=+1336.802806328" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.024661 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.024759 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.030761 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.237304 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.247855 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-4qt72"] Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.481846 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.481974 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.551050 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.981990 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerStarted","Data":"05c1fe635c1b79fed0a61153e67b08bf58566ccc0cd5071305ba2f6160b7a052"} Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.986004 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerStarted","Data":"3073890e1b6ac1ff9f663fabe29454fe5bfb09604ab7f54af106304fbdb612ef"} Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.986199 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:26 crc kubenswrapper[4989]: I1006 09:01:26.987256 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:27 crc kubenswrapper[4989]: I1006 09:01:27.013337 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" podStartSLOduration=4.013321032 podStartE2EDuration="4.013321032s" podCreationTimestamp="2025-10-06 09:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:27.002394577 +0000 UTC m=+1337.792420177" watchObservedRunningTime="2025-10-06 09:01:27.013321032 +0000 UTC m=+1337.803346612" Oct 06 09:01:27 crc kubenswrapper[4989]: I1006 09:01:27.949335 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" path="/var/lib/kubelet/pods/44b676b3-eb01-45e2-a91f-cc9b57e3e2e0/volumes" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.001491 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerStarted","Data":"fac01ec109fb091c49f98d9c013903a2e4a8fcce4fac1a9f3b69848da9287623"} Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.009470 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerStarted","Data":"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875"} Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.019289 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" event={"ID":"742c9029-45a8-47ad-9fac-cd2f6c76fc15","Type":"ContainerStarted","Data":"9f10809e8f9d1a0a063095ddee06d65c9b24be4d66c8b5f733b0a9a20eedc6ae"} Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.019475 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.023646 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerStarted","Data":"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3"} Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.037232 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" podStartSLOduration=5.037211309 podStartE2EDuration="5.037211309s" podCreationTimestamp="2025-10-06 09:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:28.03515364 +0000 UTC m=+1338.825179220" watchObservedRunningTime="2025-10-06 09:01:28.037211309 +0000 UTC m=+1338.827236889" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.667168 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:01:28 crc kubenswrapper[4989]: E1006 09:01:28.669455 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" containerName="init" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.669493 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" containerName="init" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.669712 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b676b3-eb01-45e2-a91f-cc9b57e3e2e0" containerName="init" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.670629 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.672561 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.672583 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.693844 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.818979 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lx6d\" (UniqueName: \"kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819022 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819048 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819067 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819435 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.819864 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940335 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lx6d\" (UniqueName: \"kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940562 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.940642 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.944633 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.947586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.954082 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.955888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.956919 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.960566 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.961893 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lx6d\" (UniqueName: \"kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d\") pod \"barbican-api-57c669566-xd94v\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:28 crc kubenswrapper[4989]: I1006 09:01:28.994517 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:33 crc kubenswrapper[4989]: I1006 09:01:33.574958 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:33 crc kubenswrapper[4989]: I1006 09:01:33.641925 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:01:33 crc kubenswrapper[4989]: I1006 09:01:33.642228 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="dnsmasq-dns" containerID="cri-o://341bec32385cc748dec180764963fa3e36a4420c1a18bc286a91c46965966f9e" gracePeriod=10 Oct 06 09:01:34 crc kubenswrapper[4989]: I1006 09:01:34.081141 4989 generic.go:334] "Generic (PLEG): container finished" podID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerID="341bec32385cc748dec180764963fa3e36a4420c1a18bc286a91c46965966f9e" exitCode=0 Oct 06 09:01:34 crc kubenswrapper[4989]: I1006 09:01:34.081460 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" event={"ID":"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8","Type":"ContainerDied","Data":"341bec32385cc748dec180764963fa3e36a4420c1a18bc286a91c46965966f9e"} Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.066758 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.093722 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.094147 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxvc5\" (UniqueName: \"kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.094356 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.094394 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.094419 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.094444 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0\") pod \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\" (UID: \"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8\") " Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.104605 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5" (OuterVolumeSpecName: "kube-api-access-cxvc5") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "kube-api-access-cxvc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.111394 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" event={"ID":"9cdc2aaf-9021-4371-98e5-84b2ab8f96c8","Type":"ContainerDied","Data":"bd9c2c8f221d35ab52840cf2e4f66bd35ee39ffbff9ff6af1f1592f656695b79"} Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.111445 4989 scope.go:117] "RemoveContainer" containerID="341bec32385cc748dec180764963fa3e36a4420c1a18bc286a91c46965966f9e" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.111566 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-pk6g5" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.145200 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config" (OuterVolumeSpecName: "config") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.171052 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.186374 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.197472 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxvc5\" (UniqueName: \"kubernetes.io/projected/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-kube-api-access-cxvc5\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.197492 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.197502 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.197511 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.198380 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.214587 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" (UID: "9cdc2aaf-9021-4371-98e5-84b2ab8f96c8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.253839 4989 scope.go:117] "RemoveContainer" containerID="da81eb1d67027417c132028f854ced3db667f08d0a3664d21d7c35ad4a15dd45" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.300271 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.300311 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.400015 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.503594 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.514460 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-pk6g5"] Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.527101 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.565995 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:01:35 crc kubenswrapper[4989]: E1006 09:01:35.637208 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" Oct 06 09:01:35 crc kubenswrapper[4989]: I1006 09:01:35.949503 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" path="/var/lib/kubelet/pods/9cdc2aaf-9021-4371-98e5-84b2ab8f96c8/volumes" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.125492 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerStarted","Data":"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.125671 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="ceilometer-notification-agent" containerID="cri-o://4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5" gracePeriod=30 Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.125788 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="proxy-httpd" containerID="cri-o://7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95" gracePeriod=30 Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.125799 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.125837 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="sg-core" containerID="cri-o://b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9" gracePeriod=30 Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.146894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerStarted","Data":"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.162593 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerStarted","Data":"ca7e527853be7d729e32e29acc9093bf1307493ccd1757e678492f80c2ca1c5e"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.164846 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerStarted","Data":"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.165025 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.166372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jl67c" event={"ID":"eef21e7e-a1fe-4525-8bbe-5a7adac861cb","Type":"ContainerStarted","Data":"c8539b390134394ace4899f855e06be6f8d730be158ebdbf93df2ba73394ecaa"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.169611 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerStarted","Data":"aa79f2c4a47bfd9cd6e90a260e7452c6eaabae169b34d4f1882b65c8d80ca5ac"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.169673 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerStarted","Data":"c0e526eeb3fc63bca07ba53b274454c122438ef8d646f58978c54082302aad38"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.169687 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerStarted","Data":"57d1c5d3c6d2439adf9bbdac16253bc88c648586f0f6d19ed85db5c690ee6b36"} Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.170058 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.183956 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" podStartSLOduration=11.047257373 podStartE2EDuration="14.183931374s" podCreationTimestamp="2025-10-06 09:01:22 +0000 UTC" firstStartedPulling="2025-10-06 09:01:24.502376269 +0000 UTC m=+1335.292401849" lastFinishedPulling="2025-10-06 09:01:27.63905027 +0000 UTC m=+1338.429075850" observedRunningTime="2025-10-06 09:01:36.175911523 +0000 UTC m=+1346.965937103" watchObservedRunningTime="2025-10-06 09:01:36.183931374 +0000 UTC m=+1346.973956954" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.198376 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-557b4f9dd9-blscg" podStartSLOduration=12.198347679 podStartE2EDuration="12.198347679s" podCreationTimestamp="2025-10-06 09:01:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:36.193002145 +0000 UTC m=+1346.983027745" watchObservedRunningTime="2025-10-06 09:01:36.198347679 +0000 UTC m=+1346.988373259" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.220137 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" podStartSLOduration=10.709138473 podStartE2EDuration="14.220119346s" podCreationTimestamp="2025-10-06 09:01:22 +0000 UTC" firstStartedPulling="2025-10-06 09:01:24.123950999 +0000 UTC m=+1334.913976579" lastFinishedPulling="2025-10-06 09:01:27.634931872 +0000 UTC m=+1338.424957452" observedRunningTime="2025-10-06 09:01:36.216403209 +0000 UTC m=+1347.006428819" watchObservedRunningTime="2025-10-06 09:01:36.220119346 +0000 UTC m=+1347.010144926" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.251407 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57c669566-xd94v" podStartSLOduration=8.251390386 podStartE2EDuration="8.251390386s" podCreationTimestamp="2025-10-06 09:01:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:36.248464421 +0000 UTC m=+1347.038490021" watchObservedRunningTime="2025-10-06 09:01:36.251390386 +0000 UTC m=+1347.041415966" Oct 06 09:01:36 crc kubenswrapper[4989]: I1006 09:01:36.270633 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jl67c" podStartSLOduration=2.5936392169999998 podStartE2EDuration="42.270611879s" podCreationTimestamp="2025-10-06 09:00:54 +0000 UTC" firstStartedPulling="2025-10-06 09:00:55.598339783 +0000 UTC m=+1306.388365363" lastFinishedPulling="2025-10-06 09:01:35.275312445 +0000 UTC m=+1346.065338025" observedRunningTime="2025-10-06 09:01:36.266672465 +0000 UTC m=+1347.056698055" watchObservedRunningTime="2025-10-06 09:01:36.270611879 +0000 UTC m=+1347.060637459" Oct 06 09:01:37 crc kubenswrapper[4989]: I1006 09:01:37.181007 4989 generic.go:334] "Generic (PLEG): container finished" podID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerID="7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95" exitCode=0 Oct 06 09:01:37 crc kubenswrapper[4989]: I1006 09:01:37.181334 4989 generic.go:334] "Generic (PLEG): container finished" podID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerID="b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9" exitCode=2 Oct 06 09:01:37 crc kubenswrapper[4989]: I1006 09:01:37.181057 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerDied","Data":"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95"} Oct 06 09:01:37 crc kubenswrapper[4989]: I1006 09:01:37.181547 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerDied","Data":"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9"} Oct 06 09:01:37 crc kubenswrapper[4989]: I1006 09:01:37.181628 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.214240 4989 generic.go:334] "Generic (PLEG): container finished" podID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" containerID="c8539b390134394ace4899f855e06be6f8d730be158ebdbf93df2ba73394ecaa" exitCode=0 Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.214361 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jl67c" event={"ID":"eef21e7e-a1fe-4525-8bbe-5a7adac861cb","Type":"ContainerDied","Data":"c8539b390134394ace4899f855e06be6f8d730be158ebdbf93df2ba73394ecaa"} Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.607957 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708378 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708453 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxlhh\" (UniqueName: \"kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708489 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708589 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.708811 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd\") pod \"00af6114-9242-4e7c-8ac3-416b1ff6baca\" (UID: \"00af6114-9242-4e7c-8ac3-416b1ff6baca\") " Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.709387 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.709510 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.714155 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts" (OuterVolumeSpecName: "scripts") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.715722 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh" (OuterVolumeSpecName: "kube-api-access-pxlhh") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "kube-api-access-pxlhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.735678 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.766867 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.780850 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data" (OuterVolumeSpecName: "config-data") pod "00af6114-9242-4e7c-8ac3-416b1ff6baca" (UID: "00af6114-9242-4e7c-8ac3-416b1ff6baca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810519 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxlhh\" (UniqueName: \"kubernetes.io/projected/00af6114-9242-4e7c-8ac3-416b1ff6baca-kube-api-access-pxlhh\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810564 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810581 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810597 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810612 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810630 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00af6114-9242-4e7c-8ac3-416b1ff6baca-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:40 crc kubenswrapper[4989]: I1006 09:01:40.810644 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00af6114-9242-4e7c-8ac3-416b1ff6baca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.225583 4989 generic.go:334] "Generic (PLEG): container finished" podID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerID="4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5" exitCode=0 Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.225663 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerDied","Data":"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5"} Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.225696 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00af6114-9242-4e7c-8ac3-416b1ff6baca","Type":"ContainerDied","Data":"5b0cd071ddc1ae43bfdd6d6aafd87c61599a34e700c876461142b5d579a3ac99"} Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.225716 4989 scope.go:117] "RemoveContainer" containerID="7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.226947 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.254573 4989 scope.go:117] "RemoveContainer" containerID="b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.305033 4989 scope.go:117] "RemoveContainer" containerID="4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.333401 4989 scope.go:117] "RemoveContainer" containerID="7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.333979 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95\": container with ID starting with 7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95 not found: ID does not exist" containerID="7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.334029 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95"} err="failed to get container status \"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95\": rpc error: code = NotFound desc = could not find container \"7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95\": container with ID starting with 7bb2285ce7d7bf637ca2588b971e951bef2c402ac5153b39a2cadd9b2637be95 not found: ID does not exist" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.334070 4989 scope.go:117] "RemoveContainer" containerID="b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.334472 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9\": container with ID starting with b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9 not found: ID does not exist" containerID="b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.334506 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9"} err="failed to get container status \"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9\": rpc error: code = NotFound desc = could not find container \"b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9\": container with ID starting with b4cbce6214fc55da838794a1f08afa6bb5bdbea6399c93ad082865c67400f4f9 not found: ID does not exist" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.334531 4989 scope.go:117] "RemoveContainer" containerID="4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.334968 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5\": container with ID starting with 4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5 not found: ID does not exist" containerID="4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.334998 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5"} err="failed to get container status \"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5\": rpc error: code = NotFound desc = could not find container \"4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5\": container with ID starting with 4df5b692337ad7ee772a77f48bc6092b01a4b3debb2a906ce96bb69801ece6c5 not found: ID does not exist" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.349590 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.363818 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.375920 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.376273 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="proxy-httpd" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376291 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="proxy-httpd" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.376313 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="dnsmasq-dns" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376320 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="dnsmasq-dns" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.376333 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="sg-core" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376341 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="sg-core" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.376353 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="init" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376359 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="init" Oct 06 09:01:41 crc kubenswrapper[4989]: E1006 09:01:41.376382 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="ceilometer-notification-agent" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376388 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="ceilometer-notification-agent" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376549 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="proxy-httpd" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376567 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="ceilometer-notification-agent" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376577 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cdc2aaf-9021-4371-98e5-84b2ab8f96c8" containerName="dnsmasq-dns" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.376592 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" containerName="sg-core" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.379672 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.382060 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.382689 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.401352 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.422349 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.422775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.422825 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.422878 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.422978 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.423049 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc7lf\" (UniqueName: \"kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.423081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524426 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc7lf\" (UniqueName: \"kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524539 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524556 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524628 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.524720 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.525679 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.525693 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.528836 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.529127 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.529478 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.530086 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.540114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc7lf\" (UniqueName: \"kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf\") pod \"ceilometer-0\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.616880 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jl67c" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.625961 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.626017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.626147 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.626233 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq64b\" (UniqueName: \"kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.626275 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.626332 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts\") pod \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\" (UID: \"eef21e7e-a1fe-4525-8bbe-5a7adac861cb\") " Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.628321 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.632011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.632030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b" (OuterVolumeSpecName: "kube-api-access-bq64b") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "kube-api-access-bq64b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.636433 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts" (OuterVolumeSpecName: "scripts") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.684599 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data" (OuterVolumeSpecName: "config-data") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.685948 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eef21e7e-a1fe-4525-8bbe-5a7adac861cb" (UID: "eef21e7e-a1fe-4525-8bbe-5a7adac861cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.707245 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728166 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728457 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq64b\" (UniqueName: \"kubernetes.io/projected/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-kube-api-access-bq64b\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728521 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728575 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728634 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.728713 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eef21e7e-a1fe-4525-8bbe-5a7adac861cb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:41 crc kubenswrapper[4989]: I1006 09:01:41.957404 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00af6114-9242-4e7c-8ac3-416b1ff6baca" path="/var/lib/kubelet/pods/00af6114-9242-4e7c-8ac3-416b1ff6baca/volumes" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.166231 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.247472 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jl67c" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.247517 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jl67c" event={"ID":"eef21e7e-a1fe-4525-8bbe-5a7adac861cb","Type":"ContainerDied","Data":"cbcbbeb8b5e893ba6f00b0069dba56a6c124ff055d757040209167cce4991a99"} Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.247590 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbcbbeb8b5e893ba6f00b0069dba56a6c124ff055d757040209167cce4991a99" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.252164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerStarted","Data":"3f46f378ac30bdacbb9599a2c94302cfdede1363faa0e5cd23bef863a4a5afae"} Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.499572 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:42 crc kubenswrapper[4989]: E1006 09:01:42.499985 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" containerName="cinder-db-sync" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.500003 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" containerName="cinder-db-sync" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.500196 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" containerName="cinder-db-sync" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.506912 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.512235 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.512278 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rbwjv" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.512235 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.512982 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.515908 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554113 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554183 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554223 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554331 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554360 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.554448 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhsl9\" (UniqueName: \"kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.596036 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.600131 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.609497 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.656710 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657122 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657187 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl75r\" (UniqueName: \"kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657332 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657363 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhsl9\" (UniqueName: \"kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657388 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657443 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657485 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657519 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.657548 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.663070 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.663482 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.664884 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.673243 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.680417 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhsl9\" (UniqueName: \"kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9\") pod \"cinder-scheduler-0\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.754154 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.755994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.758865 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760112 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl75r\" (UniqueName: \"kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760158 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760233 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760257 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760333 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.760373 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.761439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.762366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.762711 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.762850 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.762860 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.769638 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.785647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl75r\" (UniqueName: \"kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r\") pod \"dnsmasq-dns-5784cf869f-tjzqg\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.827235 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863738 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863794 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863840 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863894 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863912 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863938 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.863965 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h74m2\" (UniqueName: \"kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.929309 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.966626 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.966994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967041 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967048 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967159 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h74m2\" (UniqueName: \"kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967408 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967446 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967546 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.967548 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.970593 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.970790 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.972574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.973463 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:42 crc kubenswrapper[4989]: I1006 09:01:42.984113 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h74m2\" (UniqueName: \"kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2\") pod \"cinder-api-0\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " pod="openstack/cinder-api-0" Oct 06 09:01:43 crc kubenswrapper[4989]: I1006 09:01:43.081574 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:43 crc kubenswrapper[4989]: I1006 09:01:43.274459 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerStarted","Data":"28f5c762fc00ae06fb6c9144f6608fbb6331d3320538ba437fc31bf352efee77"} Oct 06 09:01:43 crc kubenswrapper[4989]: I1006 09:01:43.354020 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:43 crc kubenswrapper[4989]: I1006 09:01:43.583258 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:01:43 crc kubenswrapper[4989]: I1006 09:01:43.699374 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.288693 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerStarted","Data":"6067cea1c0acc43556fba85ff6b51dbd8502f5de46a1f8f592031fd4ddf54d15"} Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.290860 4989 generic.go:334] "Generic (PLEG): container finished" podID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerID="bc52c255a8e5d24ea77357689e284842bed6b4d5af90089348952e7a7ce407cf" exitCode=0 Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.290915 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" event={"ID":"ccacf8e5-0fb0-41bd-92de-6851964f1ffb","Type":"ContainerDied","Data":"bc52c255a8e5d24ea77357689e284842bed6b4d5af90089348952e7a7ce407cf"} Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.290940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" event={"ID":"ccacf8e5-0fb0-41bd-92de-6851964f1ffb","Type":"ContainerStarted","Data":"846a3187970586e24834545824db2dfa823583f4447875d1527e56f4827bfb51"} Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.297774 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerStarted","Data":"01256bf6a97f391803d070515b7082746ed8180fb4332c2389b95f2c7a219f39"} Oct 06 09:01:44 crc kubenswrapper[4989]: I1006 09:01:44.756037 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:45 crc kubenswrapper[4989]: I1006 09:01:45.335459 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerStarted","Data":"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5"} Oct 06 09:01:45 crc kubenswrapper[4989]: I1006 09:01:45.357591 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" event={"ID":"ccacf8e5-0fb0-41bd-92de-6851964f1ffb","Type":"ContainerStarted","Data":"27595d95b4570acc35bdf4a955f776adc14c0cb9a7b8a932b7cd599e5b47e33f"} Oct 06 09:01:45 crc kubenswrapper[4989]: I1006 09:01:45.357975 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:45 crc kubenswrapper[4989]: I1006 09:01:45.369527 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerStarted","Data":"4db33b7f195aff63b6b4a30ce90d277401a9c95f1c06a8d4618499f634e4b36a"} Oct 06 09:01:45 crc kubenswrapper[4989]: I1006 09:01:45.376408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerStarted","Data":"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573"} Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.185073 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.216317 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" podStartSLOduration=4.216299767 podStartE2EDuration="4.216299767s" podCreationTimestamp="2025-10-06 09:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:45.386965899 +0000 UTC m=+1356.176991489" watchObservedRunningTime="2025-10-06 09:01:46.216299767 +0000 UTC m=+1357.006325337" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.388960 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerStarted","Data":"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25"} Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.389255 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api-log" containerID="cri-o://972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" gracePeriod=30 Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.389438 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.389510 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api" containerID="cri-o://bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" gracePeriod=30 Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.395501 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerStarted","Data":"fae82006f4a2c91270c7f8cc8cf41044ee53d31a35ed783e907cc33fd5a5cc25"} Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.398380 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerStarted","Data":"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a"} Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.426541 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.426513937 podStartE2EDuration="4.426513937s" podCreationTimestamp="2025-10-06 09:01:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:46.418054773 +0000 UTC m=+1357.208080353" watchObservedRunningTime="2025-10-06 09:01:46.426513937 +0000 UTC m=+1357.216539627" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.869985 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.900381 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.045983146 podStartE2EDuration="4.900359944s" podCreationTimestamp="2025-10-06 09:01:42 +0000 UTC" firstStartedPulling="2025-10-06 09:01:43.360289803 +0000 UTC m=+1354.150315383" lastFinishedPulling="2025-10-06 09:01:44.214666601 +0000 UTC m=+1355.004692181" observedRunningTime="2025-10-06 09:01:46.447202432 +0000 UTC m=+1357.237228032" watchObservedRunningTime="2025-10-06 09:01:46.900359944 +0000 UTC m=+1357.690385534" Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.961278 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.962161 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api-log" containerID="cri-o://e04948b0c9de5595097016939d537f442c08815515ea97b9b0edddca18a4e910" gracePeriod=30 Oct 06 09:01:46 crc kubenswrapper[4989]: I1006 09:01:46.962671 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api" containerID="cri-o://3073890e1b6ac1ff9f663fabe29454fe5bfb09604ab7f54af106304fbdb612ef" gracePeriod=30 Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.386786 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436153 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerID="bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" exitCode=0 Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436187 4989 generic.go:334] "Generic (PLEG): container finished" podID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerID="972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" exitCode=143 Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436339 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436378 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerDied","Data":"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25"} Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436456 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerDied","Data":"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5"} Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436474 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5ad37afc-cb7f-4769-9f47-ed1f304cab24","Type":"ContainerDied","Data":"6067cea1c0acc43556fba85ff6b51dbd8502f5de46a1f8f592031fd4ddf54d15"} Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.436495 4989 scope.go:117] "RemoveContainer" containerID="bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.468045 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerStarted","Data":"e44a2fe611fc29fcee843a01096d41a72d397318445198956db5cf3782d17977"} Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.468889 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.474410 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerID="e04948b0c9de5595097016939d537f442c08815515ea97b9b0edddca18a4e910" exitCode=143 Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.475765 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerDied","Data":"e04948b0c9de5595097016939d537f442c08815515ea97b9b0edddca18a4e910"} Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.509769 4989 scope.go:117] "RemoveContainer" containerID="972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.511980 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.984721044 podStartE2EDuration="6.511957675s" podCreationTimestamp="2025-10-06 09:01:41 +0000 UTC" firstStartedPulling="2025-10-06 09:01:42.167254178 +0000 UTC m=+1352.957279758" lastFinishedPulling="2025-10-06 09:01:46.694490809 +0000 UTC m=+1357.484516389" observedRunningTime="2025-10-06 09:01:47.504380527 +0000 UTC m=+1358.294406107" watchObservedRunningTime="2025-10-06 09:01:47.511957675 +0000 UTC m=+1358.301983255" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.554716 4989 scope.go:117] "RemoveContainer" containerID="bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" Oct 06 09:01:47 crc kubenswrapper[4989]: E1006 09:01:47.557181 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25\": container with ID starting with bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25 not found: ID does not exist" containerID="bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.557284 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25"} err="failed to get container status \"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25\": rpc error: code = NotFound desc = could not find container \"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25\": container with ID starting with bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25 not found: ID does not exist" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.557373 4989 scope.go:117] "RemoveContainer" containerID="972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" Oct 06 09:01:47 crc kubenswrapper[4989]: E1006 09:01:47.557891 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5\": container with ID starting with 972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5 not found: ID does not exist" containerID="972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.558004 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5"} err="failed to get container status \"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5\": rpc error: code = NotFound desc = could not find container \"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5\": container with ID starting with 972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5 not found: ID does not exist" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.558094 4989 scope.go:117] "RemoveContainer" containerID="bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.558561 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25"} err="failed to get container status \"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25\": rpc error: code = NotFound desc = could not find container \"bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25\": container with ID starting with bdc7d542d866196b601d2ac51fb53231159727ad000ebd7569aa0c48db112e25 not found: ID does not exist" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.558676 4989 scope.go:117] "RemoveContainer" containerID="972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.559204 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5"} err="failed to get container status \"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5\": rpc error: code = NotFound desc = could not find container \"972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5\": container with ID starting with 972011b1b3b585b4d488bdabfd2275931cde6d4da2c5e25bb58c2e1ac4a047d5 not found: ID does not exist" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.587993 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.589185 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.589404 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.589811 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.589923 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h74m2\" (UniqueName: \"kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.590019 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.590098 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts\") pod \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\" (UID: \"5ad37afc-cb7f-4769-9f47-ed1f304cab24\") " Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.592030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.596468 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs" (OuterVolumeSpecName: "logs") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.598392 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.602862 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2" (OuterVolumeSpecName: "kube-api-access-h74m2") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "kube-api-access-h74m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.615857 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts" (OuterVolumeSpecName: "scripts") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.630858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.676807 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data" (OuterVolumeSpecName: "config-data") pod "5ad37afc-cb7f-4769-9f47-ed1f304cab24" (UID: "5ad37afc-cb7f-4769-9f47-ed1f304cab24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692365 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692405 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692418 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692431 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad37afc-cb7f-4769-9f47-ed1f304cab24-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692443 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h74m2\" (UniqueName: \"kubernetes.io/projected/5ad37afc-cb7f-4769-9f47-ed1f304cab24-kube-api-access-h74m2\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692455 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ad37afc-cb7f-4769-9f47-ed1f304cab24-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.692466 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad37afc-cb7f-4769-9f47-ed1f304cab24-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.809863 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.833498 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.839241 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.880829 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:47 crc kubenswrapper[4989]: E1006 09:01:47.881237 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.881254 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api" Oct 06 09:01:47 crc kubenswrapper[4989]: E1006 09:01:47.881268 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api-log" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.881275 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api-log" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.881429 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api-log" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.881449 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" containerName="cinder-api" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.882934 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.886532 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.886679 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.886732 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.891768 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895194 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895277 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895306 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895334 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd6w7\" (UniqueName: \"kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895396 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895419 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895434 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895457 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.895483 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.949488 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad37afc-cb7f-4769-9f47-ed1f304cab24" path="/var/lib/kubelet/pods/5ad37afc-cb7f-4769-9f47-ed1f304cab24/volumes" Oct 06 09:01:47 crc kubenswrapper[4989]: E1006 09:01:47.990358 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ad37afc_cb7f_4769_9f47_ed1f304cab24.slice/crio-6067cea1c0acc43556fba85ff6b51dbd8502f5de46a1f8f592031fd4ddf54d15\": RecentStats: unable to find data in memory cache]" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998157 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998249 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998287 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998311 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd6w7\" (UniqueName: \"kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998400 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998422 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:47 crc kubenswrapper[4989]: I1006 09:01:47.998445 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.002153 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.002233 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.007185 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.007295 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.013539 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.015015 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.017021 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.024052 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.026465 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd6w7\" (UniqueName: \"kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7\") pod \"cinder-api-0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.211454 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:01:48 crc kubenswrapper[4989]: I1006 09:01:48.706626 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:01:49 crc kubenswrapper[4989]: I1006 09:01:49.509899 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerStarted","Data":"61197413d388b49b459e9e0efe53d5c352fb7b325ef677d33c7410cb203309dc"} Oct 06 09:01:49 crc kubenswrapper[4989]: I1006 09:01:49.510431 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerStarted","Data":"82dc5eabc22659eeb656fbffd76032ecf98f8043f68f61542947074b4d3da8f9"} Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.414431 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:59552->10.217.0.154:9311: read: connection reset by peer" Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.415044 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:59566->10.217.0.154:9311: read: connection reset by peer" Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.522996 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerStarted","Data":"0bf97fec682ff1dbd5c988be1156a9e788940eb34663d369552ca96c1e405f14"} Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.523135 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.528183 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerID="3073890e1b6ac1ff9f663fabe29454fe5bfb09604ab7f54af106304fbdb612ef" exitCode=0 Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.528459 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerDied","Data":"3073890e1b6ac1ff9f663fabe29454fe5bfb09604ab7f54af106304fbdb612ef"} Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.549994 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.549975909 podStartE2EDuration="3.549975909s" podCreationTimestamp="2025-10-06 09:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:01:50.546453248 +0000 UTC m=+1361.336478828" watchObservedRunningTime="2025-10-06 09:01:50.549975909 +0000 UTC m=+1361.340001489" Oct 06 09:01:50 crc kubenswrapper[4989]: I1006 09:01:50.901192 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.058610 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdd26\" (UniqueName: \"kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26\") pod \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.058684 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom\") pod \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.058766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle\") pod \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.058804 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs\") pod \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.058957 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data\") pod \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\" (UID: \"a2282e6d-ba3f-434a-bace-d5e3701bbf0b\") " Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.061225 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs" (OuterVolumeSpecName: "logs") pod "a2282e6d-ba3f-434a-bace-d5e3701bbf0b" (UID: "a2282e6d-ba3f-434a-bace-d5e3701bbf0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.065613 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26" (OuterVolumeSpecName: "kube-api-access-kdd26") pod "a2282e6d-ba3f-434a-bace-d5e3701bbf0b" (UID: "a2282e6d-ba3f-434a-bace-d5e3701bbf0b"). InnerVolumeSpecName "kube-api-access-kdd26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.065971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a2282e6d-ba3f-434a-bace-d5e3701bbf0b" (UID: "a2282e6d-ba3f-434a-bace-d5e3701bbf0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.096427 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2282e6d-ba3f-434a-bace-d5e3701bbf0b" (UID: "a2282e6d-ba3f-434a-bace-d5e3701bbf0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.111671 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data" (OuterVolumeSpecName: "config-data") pod "a2282e6d-ba3f-434a-bace-d5e3701bbf0b" (UID: "a2282e6d-ba3f-434a-bace-d5e3701bbf0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.161319 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.161361 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.161374 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.161386 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdd26\" (UniqueName: \"kubernetes.io/projected/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-kube-api-access-kdd26\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.161399 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2282e6d-ba3f-434a-bace-d5e3701bbf0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.556892 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.556894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6db7d8b6f4-2xmnl" event={"ID":"a2282e6d-ba3f-434a-bace-d5e3701bbf0b","Type":"ContainerDied","Data":"d0976df43c3e6df06b0524405ab2b91644a303bb21801f327628624b1e011733"} Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.556973 4989 scope.go:117] "RemoveContainer" containerID="3073890e1b6ac1ff9f663fabe29454fe5bfb09604ab7f54af106304fbdb612ef" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.598173 4989 scope.go:117] "RemoveContainer" containerID="e04948b0c9de5595097016939d537f442c08815515ea97b9b0edddca18a4e910" Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.603175 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.610739 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6db7d8b6f4-2xmnl"] Oct 06 09:01:51 crc kubenswrapper[4989]: I1006 09:01:51.947511 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" path="/var/lib/kubelet/pods/a2282e6d-ba3f-434a-bace-d5e3701bbf0b/volumes" Oct 06 09:01:52 crc kubenswrapper[4989]: I1006 09:01:52.456214 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:52 crc kubenswrapper[4989]: I1006 09:01:52.930793 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.036234 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.036527 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="dnsmasq-dns" containerID="cri-o://9f10809e8f9d1a0a063095ddee06d65c9b24be4d66c8b5f733b0a9a20eedc6ae" gracePeriod=10 Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.095307 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.153194 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593137 4989 generic.go:334] "Generic (PLEG): container finished" podID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerID="9f10809e8f9d1a0a063095ddee06d65c9b24be4d66c8b5f733b0a9a20eedc6ae" exitCode=0 Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593343 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" event={"ID":"742c9029-45a8-47ad-9fac-cd2f6c76fc15","Type":"ContainerDied","Data":"9f10809e8f9d1a0a063095ddee06d65c9b24be4d66c8b5f733b0a9a20eedc6ae"} Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593457 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" event={"ID":"742c9029-45a8-47ad-9fac-cd2f6c76fc15","Type":"ContainerDied","Data":"6d1e677e66cae4eb8904531e607c05735d35f720e24158776bda9497370bca65"} Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593473 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d1e677e66cae4eb8904531e607c05735d35f720e24158776bda9497370bca65" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593592 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="cinder-scheduler" containerID="cri-o://dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573" gracePeriod=30 Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.593962 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="probe" containerID="cri-o://20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a" gracePeriod=30 Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.676852 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.726871 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.726932 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.726995 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.727032 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.727076 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jrgc\" (UniqueName: \"kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.727180 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config\") pod \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\" (UID: \"742c9029-45a8-47ad-9fac-cd2f6c76fc15\") " Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.749332 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc" (OuterVolumeSpecName: "kube-api-access-8jrgc") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "kube-api-access-8jrgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.786532 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.812046 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.813196 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.822591 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config" (OuterVolumeSpecName: "config") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.832822 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.832852 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.832863 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.832871 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jrgc\" (UniqueName: \"kubernetes.io/projected/742c9029-45a8-47ad-9fac-cd2f6c76fc15-kube-api-access-8jrgc\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.832889 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.833084 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "742c9029-45a8-47ad-9fac-cd2f6c76fc15" (UID: "742c9029-45a8-47ad-9fac-cd2f6c76fc15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:01:53 crc kubenswrapper[4989]: I1006 09:01:53.934617 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/742c9029-45a8-47ad-9fac-cd2f6c76fc15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:54 crc kubenswrapper[4989]: I1006 09:01:54.603818 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" Oct 06 09:01:54 crc kubenswrapper[4989]: I1006 09:01:54.634701 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:54 crc kubenswrapper[4989]: I1006 09:01:54.676476 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-dcdln"] Oct 06 09:01:54 crc kubenswrapper[4989]: I1006 09:01:54.957958 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:54 crc kubenswrapper[4989]: I1006 09:01:54.965372 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.042181 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.326190 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.398904 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.399104 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c8d7c9f64-s9jkz" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-api" containerID="cri-o://2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7" gracePeriod=30 Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.399329 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c8d7c9f64-s9jkz" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-httpd" containerID="cri-o://555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82" gracePeriod=30 Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.619067 4989 generic.go:334] "Generic (PLEG): container finished" podID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerID="20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a" exitCode=0 Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.619923 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerDied","Data":"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a"} Oct 06 09:01:55 crc kubenswrapper[4989]: I1006 09:01:55.950811 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" path="/var/lib/kubelet/pods/742c9029-45a8-47ad-9fac-cd2f6c76fc15/volumes" Oct 06 09:01:56 crc kubenswrapper[4989]: I1006 09:01:56.630241 4989 generic.go:334] "Generic (PLEG): container finished" podID="69c14157-8c35-4b71-80de-c27eb129f852" containerID="555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82" exitCode=0 Oct 06 09:01:56 crc kubenswrapper[4989]: I1006 09:01:56.630324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerDied","Data":"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82"} Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.434002 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.500900 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crz8l\" (UniqueName: \"kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l\") pod \"69c14157-8c35-4b71-80de-c27eb129f852\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.501388 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle\") pod \"69c14157-8c35-4b71-80de-c27eb129f852\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.501428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config\") pod \"69c14157-8c35-4b71-80de-c27eb129f852\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.501540 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs\") pod \"69c14157-8c35-4b71-80de-c27eb129f852\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.501623 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config\") pod \"69c14157-8c35-4b71-80de-c27eb129f852\" (UID: \"69c14157-8c35-4b71-80de-c27eb129f852\") " Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.511119 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l" (OuterVolumeSpecName: "kube-api-access-crz8l") pod "69c14157-8c35-4b71-80de-c27eb129f852" (UID: "69c14157-8c35-4b71-80de-c27eb129f852"). InnerVolumeSpecName "kube-api-access-crz8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.515967 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "69c14157-8c35-4b71-80de-c27eb129f852" (UID: "69c14157-8c35-4b71-80de-c27eb129f852"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.578740 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69c14157-8c35-4b71-80de-c27eb129f852" (UID: "69c14157-8c35-4b71-80de-c27eb129f852"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.604356 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.604396 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.604408 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crz8l\" (UniqueName: \"kubernetes.io/projected/69c14157-8c35-4b71-80de-c27eb129f852-kube-api-access-crz8l\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.612498 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config" (OuterVolumeSpecName: "config") pod "69c14157-8c35-4b71-80de-c27eb129f852" (UID: "69c14157-8c35-4b71-80de-c27eb129f852"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.616158 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "69c14157-8c35-4b71-80de-c27eb129f852" (UID: "69c14157-8c35-4b71-80de-c27eb129f852"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.641639 4989 generic.go:334] "Generic (PLEG): container finished" podID="69c14157-8c35-4b71-80de-c27eb129f852" containerID="2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7" exitCode=0 Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.641696 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerDied","Data":"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7"} Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.641720 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c8d7c9f64-s9jkz" event={"ID":"69c14157-8c35-4b71-80de-c27eb129f852","Type":"ContainerDied","Data":"30f7b6ec4c5170c507992f7f853a08947aca784f43b83a2255a4f575a7248bec"} Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.641734 4989 scope.go:117] "RemoveContainer" containerID="555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.641840 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c8d7c9f64-s9jkz" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.669980 4989 scope.go:117] "RemoveContainer" containerID="2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.689359 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.707778 4989 scope.go:117] "RemoveContainer" containerID="555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.708453 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c8d7c9f64-s9jkz"] Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.710324 4989 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.710401 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/69c14157-8c35-4b71-80de-c27eb129f852-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:57 crc kubenswrapper[4989]: E1006 09:01:57.712958 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82\": container with ID starting with 555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82 not found: ID does not exist" containerID="555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.713018 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82"} err="failed to get container status \"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82\": rpc error: code = NotFound desc = could not find container \"555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82\": container with ID starting with 555e6f60b58eb366c9b8b1d99e27c0cef54dbec3096f627acf0f390aa2ee2b82 not found: ID does not exist" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.713052 4989 scope.go:117] "RemoveContainer" containerID="2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7" Oct 06 09:01:57 crc kubenswrapper[4989]: E1006 09:01:57.713620 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7\": container with ID starting with 2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7 not found: ID does not exist" containerID="2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.713693 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7"} err="failed to get container status \"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7\": rpc error: code = NotFound desc = could not find container \"2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7\": container with ID starting with 2efa21cace3af6dbce4a5db45b0b5c065f38797c22fbc12699a65ec9a57f54c7 not found: ID does not exist" Oct 06 09:01:57 crc kubenswrapper[4989]: I1006 09:01:57.948014 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69c14157-8c35-4b71-80de-c27eb129f852" path="/var/lib/kubelet/pods/69c14157-8c35-4b71-80de-c27eb129f852/volumes" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.196850 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318280 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhsl9\" (UniqueName: \"kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318338 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318365 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318405 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318469 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318506 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle\") pod \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\" (UID: \"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e\") " Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.318516 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.319020 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.326934 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts" (OuterVolumeSpecName: "scripts") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.327824 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.329540 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9" (OuterVolumeSpecName: "kube-api-access-dhsl9") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "kube-api-access-dhsl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.383100 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.420241 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhsl9\" (UniqueName: \"kubernetes.io/projected/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-kube-api-access-dhsl9\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.420284 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.420296 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.420307 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.426349 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data" (OuterVolumeSpecName: "config-data") pod "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" (UID: "ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.521635 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.574504 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75c8ddd69c-dcdln" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: i/o timeout" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.652326 4989 generic.go:334] "Generic (PLEG): container finished" podID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerID="dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573" exitCode=0 Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.652430 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerDied","Data":"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573"} Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.652457 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e","Type":"ContainerDied","Data":"01256bf6a97f391803d070515b7082746ed8180fb4332c2389b95f2c7a219f39"} Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.652474 4989 scope.go:117] "RemoveContainer" containerID="20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.652493 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.684389 4989 scope.go:117] "RemoveContainer" containerID="dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.686812 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.698424 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.708859 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709479 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709501 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709522 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="cinder-scheduler" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709532 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="cinder-scheduler" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709543 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="dnsmasq-dns" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709552 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="dnsmasq-dns" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709572 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="probe" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709581 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="probe" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709595 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-httpd" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709602 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-httpd" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709619 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-api" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709627 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-api" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709642 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="init" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709666 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="init" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.709677 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api-log" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709685 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api-log" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709920 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api-log" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709940 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-api" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709951 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="69c14157-8c35-4b71-80de-c27eb129f852" containerName="neutron-httpd" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709964 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="742c9029-45a8-47ad-9fac-cd2f6c76fc15" containerName="dnsmasq-dns" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709978 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2282e6d-ba3f-434a-bace-d5e3701bbf0b" containerName="barbican-api" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.709991 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="cinder-scheduler" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.710009 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" containerName="probe" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.712829 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.716639 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.727017 4989 scope.go:117] "RemoveContainer" containerID="20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.727220 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.727402 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a\": container with ID starting with 20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a not found: ID does not exist" containerID="20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.727431 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a"} err="failed to get container status \"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a\": rpc error: code = NotFound desc = could not find container \"20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a\": container with ID starting with 20003d46cbdfa0cc6529ff6232ca468b397212dc88ff99ce3b834622ac80ec3a not found: ID does not exist" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.727449 4989 scope.go:117] "RemoveContainer" containerID="dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573" Oct 06 09:01:58 crc kubenswrapper[4989]: E1006 09:01:58.727955 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573\": container with ID starting with dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573 not found: ID does not exist" containerID="dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.728010 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573"} err="failed to get container status \"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573\": rpc error: code = NotFound desc = could not find container \"dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573\": container with ID starting with dd16bc54125288a2f93e044b9a9fd9f5fccc1b30f015a8c39dfdbf453c73a573 not found: ID does not exist" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739326 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739371 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739404 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739422 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739437 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.739501 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc42f\" (UniqueName: \"kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.840820 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841188 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841393 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc42f\" (UniqueName: \"kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841588 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841624 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841721 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.841825 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.844938 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.845202 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.845232 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.846276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:58 crc kubenswrapper[4989]: I1006 09:01:58.864316 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc42f\" (UniqueName: \"kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f\") pod \"cinder-scheduler-0\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " pod="openstack/cinder-scheduler-0" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.037188 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.481793 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.520113 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.521274 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.524733 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.524817 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zq9rb" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.525056 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.548202 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.656275 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.656369 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz9v\" (UniqueName: \"kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.656687 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.656712 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.667796 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerStarted","Data":"c7122261bc6367b1cee3cee833a5c28aad55d9f06208cc3a691b7d0742e6e521"} Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.736321 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: E1006 09:01:59.736959 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-mnz9v openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="4f171702-868c-4f8c-8a33-4cb5183ef36d" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.757915 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.757967 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz9v\" (UniqueName: \"kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.758021 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.758035 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.758807 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: E1006 09:01:59.760811 4989 projected.go:194] Error preparing data for projected volume kube-api-access-mnz9v for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (4f171702-868c-4f8c-8a33-4cb5183ef36d) does not match the UID in record. The object might have been deleted and then recreated Oct 06 09:01:59 crc kubenswrapper[4989]: E1006 09:01:59.760873 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v podName:4f171702-868c-4f8c-8a33-4cb5183ef36d nodeName:}" failed. No retries permitted until 2025-10-06 09:02:00.260855733 +0000 UTC m=+1371.050881313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mnz9v" (UniqueName: "kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v") pod "openstackclient" (UID: "4f171702-868c-4f8c-8a33-4cb5183ef36d") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (4f171702-868c-4f8c-8a33-4cb5183ef36d) does not match the UID in record. The object might have been deleted and then recreated Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.764982 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.765507 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.765134 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.774611 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.775792 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.823127 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.865206 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.865546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.865610 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.865688 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxn6d\" (UniqueName: \"kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.960582 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e" path="/var/lib/kubelet/pods/ffa54698-dbdd-42a8-9a2a-ad6dfd23dd9e/volumes" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.968055 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxn6d\" (UniqueName: \"kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.968125 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.968213 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.968296 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.970457 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.973946 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.976026 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:01:59 crc kubenswrapper[4989]: I1006 09:01:59.991192 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxn6d\" (UniqueName: \"kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d\") pod \"openstackclient\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " pod="openstack/openstackclient" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.145447 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.274704 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz9v\" (UniqueName: \"kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v\") pod \"openstackclient\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " pod="openstack/openstackclient" Oct 06 09:02:00 crc kubenswrapper[4989]: E1006 09:02:00.276841 4989 projected.go:194] Error preparing data for projected volume kube-api-access-mnz9v for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (4f171702-868c-4f8c-8a33-4cb5183ef36d) does not match the UID in record. The object might have been deleted and then recreated Oct 06 09:02:00 crc kubenswrapper[4989]: E1006 09:02:00.277040 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v podName:4f171702-868c-4f8c-8a33-4cb5183ef36d nodeName:}" failed. No retries permitted until 2025-10-06 09:02:01.27701806 +0000 UTC m=+1372.067043640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-mnz9v" (UniqueName: "kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v") pod "openstackclient" (UID: "4f171702-868c-4f8c-8a33-4cb5183ef36d") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (4f171702-868c-4f8c-8a33-4cb5183ef36d) does not match the UID in record. The object might have been deleted and then recreated Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.691154 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.691289 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerStarted","Data":"f4b25eb0db619fd2778bd141569f6d60f1ea4b0b8e5c38cb07786cb6a462723c"} Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.702312 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 09:02:00 crc kubenswrapper[4989]: W1006 09:02:00.702828 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod629e7fe3_c002_4313_8c5b_bdf82e4adcd5.slice/crio-47f5bdcedd2a1b358b4c465c42df38acfd521b0a20e0af48a99f3f54cefb45a3 WatchSource:0}: Error finding container 47f5bdcedd2a1b358b4c465c42df38acfd521b0a20e0af48a99f3f54cefb45a3: Status 404 returned error can't find the container with id 47f5bdcedd2a1b358b4c465c42df38acfd521b0a20e0af48a99f3f54cefb45a3 Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.712502 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.714927 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4f171702-868c-4f8c-8a33-4cb5183ef36d" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.785161 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.892496 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config\") pod \"4f171702-868c-4f8c-8a33-4cb5183ef36d\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.892574 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle\") pod \"4f171702-868c-4f8c-8a33-4cb5183ef36d\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.892626 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret\") pod \"4f171702-868c-4f8c-8a33-4cb5183ef36d\" (UID: \"4f171702-868c-4f8c-8a33-4cb5183ef36d\") " Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.893011 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnz9v\" (UniqueName: \"kubernetes.io/projected/4f171702-868c-4f8c-8a33-4cb5183ef36d-kube-api-access-mnz9v\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.893149 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4f171702-868c-4f8c-8a33-4cb5183ef36d" (UID: "4f171702-868c-4f8c-8a33-4cb5183ef36d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.901177 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f171702-868c-4f8c-8a33-4cb5183ef36d" (UID: "4f171702-868c-4f8c-8a33-4cb5183ef36d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.901212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4f171702-868c-4f8c-8a33-4cb5183ef36d" (UID: "4f171702-868c-4f8c-8a33-4cb5183ef36d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.995441 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.995723 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:00 crc kubenswrapper[4989]: I1006 09:02:00.995735 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4f171702-868c-4f8c-8a33-4cb5183ef36d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.701853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"629e7fe3-c002-4313-8c5b-bdf82e4adcd5","Type":"ContainerStarted","Data":"47f5bdcedd2a1b358b4c465c42df38acfd521b0a20e0af48a99f3f54cefb45a3"} Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.705678 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerStarted","Data":"e2d916bd8ce663a1374a2102f14126cf7f39b34e74293d78b95861c7f07ea483"} Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.705712 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.729539 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.729520371 podStartE2EDuration="3.729520371s" podCreationTimestamp="2025-10-06 09:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:01.723735424 +0000 UTC m=+1372.513761024" watchObservedRunningTime="2025-10-06 09:02:01.729520371 +0000 UTC m=+1372.519545951" Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.732286 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="4f171702-868c-4f8c-8a33-4cb5183ef36d" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" Oct 06 09:02:01 crc kubenswrapper[4989]: I1006 09:02:01.950431 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f171702-868c-4f8c-8a33-4cb5183ef36d" path="/var/lib/kubelet/pods/4f171702-868c-4f8c-8a33-4cb5183ef36d/volumes" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.789419 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.791387 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.800422 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.802007 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.802263 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.803395 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.952619 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.952822 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.952847 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.952867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.952942 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.953187 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j95dt\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.953302 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:03 crc kubenswrapper[4989]: I1006 09:02:03.953392 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.038072 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055278 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055361 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j95dt\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055409 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055528 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.055547 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.056530 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.057647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.067376 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.067748 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.068770 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.083144 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j95dt\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.092306 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.098909 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs\") pod \"swift-proxy-64d9fcfdcc-xfhh5\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.124320 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.519528 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.520132 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-central-agent" containerID="cri-o://28f5c762fc00ae06fb6c9144f6608fbb6331d3320538ba437fc31bf352efee77" gracePeriod=30 Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.520179 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="sg-core" containerID="cri-o://fae82006f4a2c91270c7f8cc8cf41044ee53d31a35ed783e907cc33fd5a5cc25" gracePeriod=30 Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.520250 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-notification-agent" containerID="cri-o://4db33b7f195aff63b6b4a30ce90d277401a9c95f1c06a8d4618499f634e4b36a" gracePeriod=30 Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.520291 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="proxy-httpd" containerID="cri-o://e44a2fe611fc29fcee843a01096d41a72d397318445198956db5cf3782d17977" gracePeriod=30 Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.530600 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.720615 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.750942 4989 generic.go:334] "Generic (PLEG): container finished" podID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerID="fae82006f4a2c91270c7f8cc8cf41044ee53d31a35ed783e907cc33fd5a5cc25" exitCode=2 Oct 06 09:02:04 crc kubenswrapper[4989]: I1006 09:02:04.751020 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerDied","Data":"fae82006f4a2c91270c7f8cc8cf41044ee53d31a35ed783e907cc33fd5a5cc25"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.763559 4989 generic.go:334] "Generic (PLEG): container finished" podID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerID="e44a2fe611fc29fcee843a01096d41a72d397318445198956db5cf3782d17977" exitCode=0 Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.764126 4989 generic.go:334] "Generic (PLEG): container finished" podID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerID="28f5c762fc00ae06fb6c9144f6608fbb6331d3320538ba437fc31bf352efee77" exitCode=0 Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.763629 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerDied","Data":"e44a2fe611fc29fcee843a01096d41a72d397318445198956db5cf3782d17977"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.764167 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerDied","Data":"28f5c762fc00ae06fb6c9144f6608fbb6331d3320538ba437fc31bf352efee77"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.769560 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerStarted","Data":"24bd4b9578823a45ed78764215744ad5b434e9e908686d6d508220f295123fc5"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.769602 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerStarted","Data":"4368459492d84638518335c2551dd210d155262a4a5fb3137e085d2b4dab20cc"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.769617 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerStarted","Data":"8d21c1f0abb03bb68ac6359477e03082b1bf2546df6cdaadd81935a06b5a5193"} Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.769698 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:05 crc kubenswrapper[4989]: I1006 09:02:05.804247 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" podStartSLOduration=2.804222836 podStartE2EDuration="2.804222836s" podCreationTimestamp="2025-10-06 09:02:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:05.794759523 +0000 UTC m=+1376.584785103" watchObservedRunningTime="2025-10-06 09:02:05.804222836 +0000 UTC m=+1376.594248416" Oct 06 09:02:06 crc kubenswrapper[4989]: I1006 09:02:06.779356 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:08 crc kubenswrapper[4989]: I1006 09:02:08.811132 4989 generic.go:334] "Generic (PLEG): container finished" podID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerID="4db33b7f195aff63b6b4a30ce90d277401a9c95f1c06a8d4618499f634e4b36a" exitCode=0 Oct 06 09:02:08 crc kubenswrapper[4989]: I1006 09:02:08.811282 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerDied","Data":"4db33b7f195aff63b6b4a30ce90d277401a9c95f1c06a8d4618499f634e4b36a"} Oct 06 09:02:09 crc kubenswrapper[4989]: I1006 09:02:09.141829 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:09 crc kubenswrapper[4989]: I1006 09:02:09.301436 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 09:02:10 crc kubenswrapper[4989]: I1006 09:02:10.222007 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:10 crc kubenswrapper[4989]: I1006 09:02:10.222199 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="55218d49-6773-44cc-a7c0-38ac71f8f66d" containerName="kube-state-metrics" containerID="cri-o://354435a487e4639127dcd86ea2b36527b64d79471e09035787232d88307f80f1" gracePeriod=30 Oct 06 09:02:10 crc kubenswrapper[4989]: I1006 09:02:10.838597 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55218d49-6773-44cc-a7c0-38ac71f8f66d","Type":"ContainerDied","Data":"354435a487e4639127dcd86ea2b36527b64d79471e09035787232d88307f80f1"} Oct 06 09:02:10 crc kubenswrapper[4989]: I1006 09:02:10.838534 4989 generic.go:334] "Generic (PLEG): container finished" podID="55218d49-6773-44cc-a7c0-38ac71f8f66d" containerID="354435a487e4639127dcd86ea2b36527b64d79471e09035787232d88307f80f1" exitCode=2 Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.517555 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.569750 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.626276 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9mvw\" (UniqueName: \"kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw\") pod \"55218d49-6773-44cc-a7c0-38ac71f8f66d\" (UID: \"55218d49-6773-44cc-a7c0-38ac71f8f66d\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.632860 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw" (OuterVolumeSpecName: "kube-api-access-n9mvw") pod "55218d49-6773-44cc-a7c0-38ac71f8f66d" (UID: "55218d49-6773-44cc-a7c0-38ac71f8f66d"). InnerVolumeSpecName "kube-api-access-n9mvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727487 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727576 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727640 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727716 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727798 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc7lf\" (UniqueName: \"kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727878 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.727919 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd\") pod \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\" (UID: \"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24\") " Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.728130 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.728438 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.729962 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.729996 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9mvw\" (UniqueName: \"kubernetes.io/projected/55218d49-6773-44cc-a7c0-38ac71f8f66d-kube-api-access-n9mvw\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.730011 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.745968 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-jfdd5"] Oct 06 09:02:11 crc kubenswrapper[4989]: E1006 09:02:11.746425 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-notification-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746443 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-notification-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: E1006 09:02:11.746459 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55218d49-6773-44cc-a7c0-38ac71f8f66d" containerName="kube-state-metrics" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746465 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="55218d49-6773-44cc-a7c0-38ac71f8f66d" containerName="kube-state-metrics" Oct 06 09:02:11 crc kubenswrapper[4989]: E1006 09:02:11.746488 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="sg-core" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746494 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="sg-core" Oct 06 09:02:11 crc kubenswrapper[4989]: E1006 09:02:11.746508 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="proxy-httpd" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746515 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="proxy-httpd" Oct 06 09:02:11 crc kubenswrapper[4989]: E1006 09:02:11.746526 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-central-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746532 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-central-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746720 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="55218d49-6773-44cc-a7c0-38ac71f8f66d" containerName="kube-state-metrics" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746739 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="sg-core" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746755 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-central-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746764 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="proxy-httpd" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.746775 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" containerName="ceilometer-notification-agent" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.747361 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.750534 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf" (OuterVolumeSpecName: "kube-api-access-fc7lf") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "kube-api-access-fc7lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.759389 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts" (OuterVolumeSpecName: "scripts") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.768487 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jfdd5"] Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.773793 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.827958 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.836264 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf7pt\" (UniqueName: \"kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt\") pod \"nova-api-db-create-jfdd5\" (UID: \"b65de4a0-67b5-446e-896a-6934b9d11e9b\") " pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.836439 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.836455 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.836468 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc7lf\" (UniqueName: \"kubernetes.io/projected/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-kube-api-access-fc7lf\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.836481 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.848409 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-9mh59"] Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.849832 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.859462 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data" (OuterVolumeSpecName: "config-data") pod "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" (UID: "f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.894299 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"629e7fe3-c002-4313-8c5b-bdf82e4adcd5","Type":"ContainerStarted","Data":"ca2377b7cd7f73d36c389bf39d5a5b793d4492f8b024e80813557074b081fb4c"} Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.902572 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9mh59"] Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.908597 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"55218d49-6773-44cc-a7c0-38ac71f8f66d","Type":"ContainerDied","Data":"9be20d61f36b4f7d6aa2aec9cfbd9c61805a60775e082b38cfcb0fbcadcf56f2"} Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.908646 4989 scope.go:117] "RemoveContainer" containerID="354435a487e4639127dcd86ea2b36527b64d79471e09035787232d88307f80f1" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.908784 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.923364 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.388523687 podStartE2EDuration="12.923347014s" podCreationTimestamp="2025-10-06 09:01:59 +0000 UTC" firstStartedPulling="2025-10-06 09:02:00.706075934 +0000 UTC m=+1371.496101514" lastFinishedPulling="2025-10-06 09:02:11.240899261 +0000 UTC m=+1382.030924841" observedRunningTime="2025-10-06 09:02:11.920290535 +0000 UTC m=+1382.710316135" watchObservedRunningTime="2025-10-06 09:02:11.923347014 +0000 UTC m=+1382.713372594" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.925125 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24","Type":"ContainerDied","Data":"3f46f378ac30bdacbb9599a2c94302cfdede1363faa0e5cd23bef863a4a5afae"} Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.925229 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.939977 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf7pt\" (UniqueName: \"kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt\") pod \"nova-api-db-create-jfdd5\" (UID: \"b65de4a0-67b5-446e-896a-6934b9d11e9b\") " pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.940334 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.976935 4989 scope.go:117] "RemoveContainer" containerID="e44a2fe611fc29fcee843a01096d41a72d397318445198956db5cf3782d17977" Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.979017 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:11 crc kubenswrapper[4989]: I1006 09:02:11.979151 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf7pt\" (UniqueName: \"kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt\") pod \"nova-api-db-create-jfdd5\" (UID: \"b65de4a0-67b5-446e-896a-6934b9d11e9b\") " pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.004739 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5fmhh"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.006449 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.030846 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.041430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62hv\" (UniqueName: \"kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv\") pod \"nova-cell0-db-create-9mh59\" (UID: \"86625194-ed32-446c-869a-d1a8d30a5c93\") " pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.042932 4989 scope.go:117] "RemoveContainer" containerID="fae82006f4a2c91270c7f8cc8cf41044ee53d31a35ed783e907cc33fd5a5cc25" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.062737 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5fmhh"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.068005 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.077725 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.082427 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.084069 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.089358 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.089456 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.089802 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-99llx" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.100884 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.136284 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.142874 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62hv\" (UniqueName: \"kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv\") pod \"nova-cell0-db-create-9mh59\" (UID: \"86625194-ed32-446c-869a-d1a8d30a5c93\") " pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.143010 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzd4h\" (UniqueName: \"kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h\") pod \"nova-cell1-db-create-5fmhh\" (UID: \"d81dcb4a-9801-4a2d-a106-0b7916a6b569\") " pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.157118 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.160239 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.166193 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.166462 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.182022 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62hv\" (UniqueName: \"kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv\") pod \"nova-cell0-db-create-9mh59\" (UID: \"86625194-ed32-446c-869a-d1a8d30a5c93\") " pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.198342 4989 scope.go:117] "RemoveContainer" containerID="4db33b7f195aff63b6b4a30ce90d277401a9c95f1c06a8d4618499f634e4b36a" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.209887 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.210065 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.244917 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj2bs\" (UniqueName: \"kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.244972 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245034 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245062 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzd4h\" (UniqueName: \"kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h\") pod \"nova-cell1-db-create-5fmhh\" (UID: \"d81dcb4a-9801-4a2d-a106-0b7916a6b569\") " pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245082 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245100 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245131 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245146 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqh8\" (UniqueName: \"kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245171 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245187 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245213 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.245233 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.263871 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzd4h\" (UniqueName: \"kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h\") pod \"nova-cell1-db-create-5fmhh\" (UID: \"d81dcb4a-9801-4a2d-a106-0b7916a6b569\") " pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.331519 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350265 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350316 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350351 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj2bs\" (UniqueName: \"kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350522 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350602 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350637 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350674 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350714 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350733 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqh8\" (UniqueName: \"kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.350768 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.354476 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.354704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.358734 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.359499 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.360946 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.366526 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.368581 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.368781 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.369221 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.372481 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqh8\" (UniqueName: \"kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8\") pod \"ceilometer-0\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.375520 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj2bs\" (UniqueName: \"kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs\") pod \"kube-state-metrics-0\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.388249 4989 scope.go:117] "RemoveContainer" containerID="28f5c762fc00ae06fb6c9144f6608fbb6331d3320538ba437fc31bf352efee77" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.430676 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.478828 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.708598 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jfdd5"] Oct 06 09:02:12 crc kubenswrapper[4989]: W1006 09:02:12.709792 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb65de4a0_67b5_446e_896a_6934b9d11e9b.slice/crio-aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507 WatchSource:0}: Error finding container aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507: Status 404 returned error can't find the container with id aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507 Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.748236 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.843840 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9mh59"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.945801 4989 generic.go:334] "Generic (PLEG): container finished" podID="b65de4a0-67b5-446e-896a-6934b9d11e9b" containerID="1099b3c6e3fd133aabba584c4cf9b090c37f9c1f70d86e0fdfad871f4f32cd21" exitCode=0 Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.946291 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jfdd5" event={"ID":"b65de4a0-67b5-446e-896a-6934b9d11e9b","Type":"ContainerDied","Data":"1099b3c6e3fd133aabba584c4cf9b090c37f9c1f70d86e0fdfad871f4f32cd21"} Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.946331 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jfdd5" event={"ID":"b65de4a0-67b5-446e-896a-6934b9d11e9b","Type":"ContainerStarted","Data":"aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507"} Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.955264 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5fmhh"] Oct 06 09:02:12 crc kubenswrapper[4989]: I1006 09:02:12.955506 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9mh59" event={"ID":"86625194-ed32-446c-869a-d1a8d30a5c93","Type":"ContainerStarted","Data":"f6f2f52e67baf7a5161a2d0e6dffd2bfab97e67907b70f9f1e2d43dd850c7aa4"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.076709 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.083906 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:13 crc kubenswrapper[4989]: W1006 09:02:13.103771 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef8cd741_ab58_4662_872c_8955c47ef9ca.slice/crio-94377efccbefe338de524e9a0f206ca4ee08411cad63b480db8cc334e837a262 WatchSource:0}: Error finding container 94377efccbefe338de524e9a0f206ca4ee08411cad63b480db8cc334e837a262: Status 404 returned error can't find the container with id 94377efccbefe338de524e9a0f206ca4ee08411cad63b480db8cc334e837a262 Oct 06 09:02:13 crc kubenswrapper[4989]: W1006 09:02:13.104359 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c5263b9_2c57_4f18_86b6_864cce8af85c.slice/crio-8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59 WatchSource:0}: Error finding container 8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59: Status 404 returned error can't find the container with id 8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.106486 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.642363 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.642879 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-log" containerID="cri-o://374465b29904cc9b24031dc73ee4754d240a2e11bccbf776e34fa10c9f5ff9fc" gracePeriod=30 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.642971 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-httpd" containerID="cri-o://db767707a8ce13eb95542aa72cf874e036f3e2c48fc44f5cc7fd81804478b34b" gracePeriod=30 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.950898 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55218d49-6773-44cc-a7c0-38ac71f8f66d" path="/var/lib/kubelet/pods/55218d49-6773-44cc-a7c0-38ac71f8f66d/volumes" Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.952107 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24" path="/var/lib/kubelet/pods/f9c63ec9-c370-4c4c-b9f7-c6f9c6f38e24/volumes" Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.968079 4989 generic.go:334] "Generic (PLEG): container finished" podID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerID="374465b29904cc9b24031dc73ee4754d240a2e11bccbf776e34fa10c9f5ff9fc" exitCode=143 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.968178 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerDied","Data":"374465b29904cc9b24031dc73ee4754d240a2e11bccbf776e34fa10c9f5ff9fc"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.970524 4989 generic.go:334] "Generic (PLEG): container finished" podID="86625194-ed32-446c-869a-d1a8d30a5c93" containerID="aef057b8eb8304f36a1d6725c1a189e2b948c62649b59415ad9742a99c463eb1" exitCode=0 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.970660 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9mh59" event={"ID":"86625194-ed32-446c-869a-d1a8d30a5c93","Type":"ContainerDied","Data":"aef057b8eb8304f36a1d6725c1a189e2b948c62649b59415ad9742a99c463eb1"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.972800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c5263b9-2c57-4f18-86b6-864cce8af85c","Type":"ContainerStarted","Data":"6e66f1425bf72188ce8811a749a7167ac2b29573e0f8e328426cf8a51343c25d"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.972829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c5263b9-2c57-4f18-86b6-864cce8af85c","Type":"ContainerStarted","Data":"8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.972925 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.974845 4989 generic.go:334] "Generic (PLEG): container finished" podID="d81dcb4a-9801-4a2d-a106-0b7916a6b569" containerID="4f5bd568ffa89c45154d075061eda3a39f78f1665fe2c860e6ff950b52542672" exitCode=0 Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.974891 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5fmhh" event={"ID":"d81dcb4a-9801-4a2d-a106-0b7916a6b569","Type":"ContainerDied","Data":"4f5bd568ffa89c45154d075061eda3a39f78f1665fe2c860e6ff950b52542672"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.974939 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5fmhh" event={"ID":"d81dcb4a-9801-4a2d-a106-0b7916a6b569","Type":"ContainerStarted","Data":"45390f5c06e9b0521243722a6cbf50c0c00df757fb51ae4070248cbaeb01f207"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.977236 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerStarted","Data":"1280c170a12709fa4a2cda4733cc078a2937e3460f0d2fb63d5c9b83282fee7f"} Oct 06 09:02:13 crc kubenswrapper[4989]: I1006 09:02:13.977278 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerStarted","Data":"94377efccbefe338de524e9a0f206ca4ee08411cad63b480db8cc334e837a262"} Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.017573 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.6724650690000002 podStartE2EDuration="3.017555371s" podCreationTimestamp="2025-10-06 09:02:11 +0000 UTC" firstStartedPulling="2025-10-06 09:02:13.108868244 +0000 UTC m=+1383.898893824" lastFinishedPulling="2025-10-06 09:02:13.453958546 +0000 UTC m=+1384.243984126" observedRunningTime="2025-10-06 09:02:14.014256236 +0000 UTC m=+1384.804281816" watchObservedRunningTime="2025-10-06 09:02:14.017555371 +0000 UTC m=+1384.807580951" Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.137293 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.371747 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.493823 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf7pt\" (UniqueName: \"kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt\") pod \"b65de4a0-67b5-446e-896a-6934b9d11e9b\" (UID: \"b65de4a0-67b5-446e-896a-6934b9d11e9b\") " Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.498354 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt" (OuterVolumeSpecName: "kube-api-access-vf7pt") pod "b65de4a0-67b5-446e-896a-6934b9d11e9b" (UID: "b65de4a0-67b5-446e-896a-6934b9d11e9b"). InnerVolumeSpecName "kube-api-access-vf7pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:14 crc kubenswrapper[4989]: I1006 09:02:14.595836 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf7pt\" (UniqueName: \"kubernetes.io/projected/b65de4a0-67b5-446e-896a-6934b9d11e9b-kube-api-access-vf7pt\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.000227 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerStarted","Data":"40f66117b81ed15de75788b89b54374fee5f59e5999963b377716605187461ed"} Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.016617 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jfdd5" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.018575 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jfdd5" event={"ID":"b65de4a0-67b5-446e-896a-6934b9d11e9b","Type":"ContainerDied","Data":"aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507"} Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.018612 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aac33e3199b18072d251d8436e95b9f160d9907a4a7bffb995d719a89d77e507" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.413624 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.517336 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzd4h\" (UniqueName: \"kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h\") pod \"d81dcb4a-9801-4a2d-a106-0b7916a6b569\" (UID: \"d81dcb4a-9801-4a2d-a106-0b7916a6b569\") " Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.521891 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h" (OuterVolumeSpecName: "kube-api-access-mzd4h") pod "d81dcb4a-9801-4a2d-a106-0b7916a6b569" (UID: "d81dcb4a-9801-4a2d-a106-0b7916a6b569"). InnerVolumeSpecName "kube-api-access-mzd4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.545769 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.619446 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzd4h\" (UniqueName: \"kubernetes.io/projected/d81dcb4a-9801-4a2d-a106-0b7916a6b569-kube-api-access-mzd4h\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.720969 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d62hv\" (UniqueName: \"kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv\") pod \"86625194-ed32-446c-869a-d1a8d30a5c93\" (UID: \"86625194-ed32-446c-869a-d1a8d30a5c93\") " Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.725397 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv" (OuterVolumeSpecName: "kube-api-access-d62hv") pod "86625194-ed32-446c-869a-d1a8d30a5c93" (UID: "86625194-ed32-446c-869a-d1a8d30a5c93"). InnerVolumeSpecName "kube-api-access-d62hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:15 crc kubenswrapper[4989]: I1006 09:02:15.823074 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d62hv\" (UniqueName: \"kubernetes.io/projected/86625194-ed32-446c-869a-d1a8d30a5c93-kube-api-access-d62hv\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.034220 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5fmhh" Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.034235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5fmhh" event={"ID":"d81dcb4a-9801-4a2d-a106-0b7916a6b569","Type":"ContainerDied","Data":"45390f5c06e9b0521243722a6cbf50c0c00df757fb51ae4070248cbaeb01f207"} Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.034276 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45390f5c06e9b0521243722a6cbf50c0c00df757fb51ae4070248cbaeb01f207" Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.037339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerStarted","Data":"96fed2b62148d2ee8435c799fb4d5985da17893469900df4534055b6130b75c1"} Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.039105 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9mh59" event={"ID":"86625194-ed32-446c-869a-d1a8d30a5c93","Type":"ContainerDied","Data":"f6f2f52e67baf7a5161a2d0e6dffd2bfab97e67907b70f9f1e2d43dd850c7aa4"} Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.039126 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6f2f52e67baf7a5161a2d0e6dffd2bfab97e67907b70f9f1e2d43dd850c7aa4" Oct 06 09:02:16 crc kubenswrapper[4989]: I1006 09:02:16.039171 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9mh59" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.050436 4989 generic.go:334] "Generic (PLEG): container finished" podID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerID="db767707a8ce13eb95542aa72cf874e036f3e2c48fc44f5cc7fd81804478b34b" exitCode=0 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.050519 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerDied","Data":"db767707a8ce13eb95542aa72cf874e036f3e2c48fc44f5cc7fd81804478b34b"} Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.053023 4989 generic.go:334] "Generic (PLEG): container finished" podID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerID="3061bb1ec9fc4bfdf9250936563634f747e72909374e0d922e1b15d16d70592b" exitCode=1 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.053047 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerDied","Data":"3061bb1ec9fc4bfdf9250936563634f747e72909374e0d922e1b15d16d70592b"} Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.053184 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-central-agent" containerID="cri-o://1280c170a12709fa4a2cda4733cc078a2937e3460f0d2fb63d5c9b83282fee7f" gracePeriod=30 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.053640 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="sg-core" containerID="cri-o://96fed2b62148d2ee8435c799fb4d5985da17893469900df4534055b6130b75c1" gracePeriod=30 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.053724 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-notification-agent" containerID="cri-o://40f66117b81ed15de75788b89b54374fee5f59e5999963b377716605187461ed" gracePeriod=30 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.339177 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454164 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454232 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454279 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454410 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454549 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454590 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltvw8\" (UniqueName: \"kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454706 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.454776 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"8817e72b-0ee6-4c49-973d-4d5de2baf242\" (UID: \"8817e72b-0ee6-4c49-973d-4d5de2baf242\") " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.455271 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.455500 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.455782 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs" (OuterVolumeSpecName: "logs") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.461061 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.461608 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8" (OuterVolumeSpecName: "kube-api-access-ltvw8") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "kube-api-access-ltvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.462170 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts" (OuterVolumeSpecName: "scripts") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.493300 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.519683 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.549112 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data" (OuterVolumeSpecName: "config-data") pod "8817e72b-0ee6-4c49-973d-4d5de2baf242" (UID: "8817e72b-0ee6-4c49-973d-4d5de2baf242"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557784 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8817e72b-0ee6-4c49-973d-4d5de2baf242-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557828 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557838 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557847 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557859 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltvw8\" (UniqueName: \"kubernetes.io/projected/8817e72b-0ee6-4c49-973d-4d5de2baf242-kube-api-access-ltvw8\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557872 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8817e72b-0ee6-4c49-973d-4d5de2baf242-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.557912 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.577612 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.597589 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.598104 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-log" containerID="cri-o://7d67e6989f2726f5fff92491c13e5646677be4b188c274a06e93bde0e31ed8f4" gracePeriod=30 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.598922 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-httpd" containerID="cri-o://6b63e8b67a9d19855d3c01aa43cd7cfb864abb3738572d14b99446cf360b4da7" gracePeriod=30 Oct 06 09:02:17 crc kubenswrapper[4989]: I1006 09:02:17.659296 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.064841 4989 generic.go:334] "Generic (PLEG): container finished" podID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerID="7d67e6989f2726f5fff92491c13e5646677be4b188c274a06e93bde0e31ed8f4" exitCode=143 Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.064975 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerDied","Data":"7d67e6989f2726f5fff92491c13e5646677be4b188c274a06e93bde0e31ed8f4"} Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.068607 4989 generic.go:334] "Generic (PLEG): container finished" podID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerID="96fed2b62148d2ee8435c799fb4d5985da17893469900df4534055b6130b75c1" exitCode=2 Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.068639 4989 generic.go:334] "Generic (PLEG): container finished" podID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerID="40f66117b81ed15de75788b89b54374fee5f59e5999963b377716605187461ed" exitCode=0 Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.068709 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerDied","Data":"96fed2b62148d2ee8435c799fb4d5985da17893469900df4534055b6130b75c1"} Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.068737 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerDied","Data":"40f66117b81ed15de75788b89b54374fee5f59e5999963b377716605187461ed"} Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.071242 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8817e72b-0ee6-4c49-973d-4d5de2baf242","Type":"ContainerDied","Data":"e87e111fe66777464f64e3cdb17073626edd1338f141742efb09025d841a4215"} Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.071467 4989 scope.go:117] "RemoveContainer" containerID="db767707a8ce13eb95542aa72cf874e036f3e2c48fc44f5cc7fd81804478b34b" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.071966 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.107906 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.121932 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.141731 4989 scope.go:117] "RemoveContainer" containerID="374465b29904cc9b24031dc73ee4754d240a2e11bccbf776e34fa10c9f5ff9fc" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145519 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:18 crc kubenswrapper[4989]: E1006 09:02:18.145862 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-httpd" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145873 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-httpd" Oct 06 09:02:18 crc kubenswrapper[4989]: E1006 09:02:18.145892 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d81dcb4a-9801-4a2d-a106-0b7916a6b569" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145898 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d81dcb4a-9801-4a2d-a106-0b7916a6b569" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: E1006 09:02:18.145914 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65de4a0-67b5-446e-896a-6934b9d11e9b" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145920 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65de4a0-67b5-446e-896a-6934b9d11e9b" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: E1006 09:02:18.145926 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86625194-ed32-446c-869a-d1a8d30a5c93" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145931 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="86625194-ed32-446c-869a-d1a8d30a5c93" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: E1006 09:02:18.145944 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-log" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.145949 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-log" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.146115 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b65de4a0-67b5-446e-896a-6934b9d11e9b" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.146126 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-httpd" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.146143 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d81dcb4a-9801-4a2d-a106-0b7916a6b569" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.146152 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="86625194-ed32-446c-869a-d1a8d30a5c93" containerName="mariadb-database-create" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.146160 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" containerName="glance-log" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.147302 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.149572 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.149794 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.156098 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.270257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77fnz\" (UniqueName: \"kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.270779 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.270939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.271212 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.271238 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.271272 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.271304 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.271465 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372544 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77fnz\" (UniqueName: \"kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372731 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372755 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372778 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372823 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372846 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.372877 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.373289 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.374645 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.377179 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.382697 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.384971 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.385375 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.386873 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.401313 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77fnz\" (UniqueName: \"kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.412313 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " pod="openstack/glance-default-external-api-0" Oct 06 09:02:18 crc kubenswrapper[4989]: I1006 09:02:18.478524 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:02:19 crc kubenswrapper[4989]: I1006 09:02:19.076181 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:02:19 crc kubenswrapper[4989]: W1006 09:02:19.081875 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74245807_f752_4209_a489_d3e746b5e1fb.slice/crio-895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a WatchSource:0}: Error finding container 895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a: Status 404 returned error can't find the container with id 895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a Oct 06 09:02:19 crc kubenswrapper[4989]: I1006 09:02:19.950820 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8817e72b-0ee6-4c49-973d-4d5de2baf242" path="/var/lib/kubelet/pods/8817e72b-0ee6-4c49-973d-4d5de2baf242/volumes" Oct 06 09:02:20 crc kubenswrapper[4989]: I1006 09:02:20.096236 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerStarted","Data":"1503150658fd3c0f67b76860cf73488324464b757cbc4426f8c36b6c4913eb11"} Oct 06 09:02:20 crc kubenswrapper[4989]: I1006 09:02:20.096280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerStarted","Data":"895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a"} Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.114225 4989 generic.go:334] "Generic (PLEG): container finished" podID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerID="6b63e8b67a9d19855d3c01aa43cd7cfb864abb3738572d14b99446cf360b4da7" exitCode=0 Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.114392 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerDied","Data":"6b63e8b67a9d19855d3c01aa43cd7cfb864abb3738572d14b99446cf360b4da7"} Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.120980 4989 generic.go:334] "Generic (PLEG): container finished" podID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerID="1280c170a12709fa4a2cda4733cc078a2937e3460f0d2fb63d5c9b83282fee7f" exitCode=0 Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.121029 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerDied","Data":"1280c170a12709fa4a2cda4733cc078a2937e3460f0d2fb63d5c9b83282fee7f"} Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.123345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerStarted","Data":"9e3f96ebea5bce42b75defb3a323c9cc21c5adae1bd4c054f3fd804d7a885721"} Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.168043 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.168021722 podStartE2EDuration="3.168021722s" podCreationTimestamp="2025-10-06 09:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:21.157972232 +0000 UTC m=+1391.947997812" watchObservedRunningTime="2025-10-06 09:02:21.168021722 +0000 UTC m=+1391.958047302" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.357533 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.365504 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435408 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435523 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435564 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435597 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-884pl\" (UniqueName: \"kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435638 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435724 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435785 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435843 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsqh8\" (UniqueName: \"kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435884 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435907 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435930 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435952 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.435976 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle\") pod \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\" (UID: \"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.436036 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd\") pod \"ef8cd741-ab58-4662-872c-8955c47ef9ca\" (UID: \"ef8cd741-ab58-4662-872c-8955c47ef9ca\") " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.436248 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.436524 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.436879 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs" (OuterVolumeSpecName: "logs") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.436898 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.437309 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.437704 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ef8cd741-ab58-4662-872c-8955c47ef9ca-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.445227 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts" (OuterVolumeSpecName: "scripts") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.445396 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts" (OuterVolumeSpecName: "scripts") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.446913 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl" (OuterVolumeSpecName: "kube-api-access-884pl") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "kube-api-access-884pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.448802 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8" (OuterVolumeSpecName: "kube-api-access-zsqh8") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "kube-api-access-zsqh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.449012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.478408 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.490836 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.518144 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.520783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data" (OuterVolumeSpecName: "config-data") pod "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" (UID: "1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.537495 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539634 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539699 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-884pl\" (UniqueName: \"kubernetes.io/projected/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-kube-api-access-884pl\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539715 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539725 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539735 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539769 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539783 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsqh8\" (UniqueName: \"kubernetes.io/projected/ef8cd741-ab58-4662-872c-8955c47ef9ca-kube-api-access-zsqh8\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539793 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539803 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539813 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539823 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.539833 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.562957 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.574977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data" (OuterVolumeSpecName: "config-data") pod "ef8cd741-ab58-4662-872c-8955c47ef9ca" (UID: "ef8cd741-ab58-4662-872c-8955c47ef9ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.641564 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.641595 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef8cd741-ab58-4662-872c-8955c47ef9ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.983907 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-ded3-account-create-2p5m9"] Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984390 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-central-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984417 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-central-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984439 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-log" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984448 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-log" Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984459 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="proxy-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984466 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="proxy-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984486 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="sg-core" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984493 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="sg-core" Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984504 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984510 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: E1006 09:02:21.984528 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-notification-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984534 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-notification-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984777 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-central-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984797 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984809 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" containerName="glance-log" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984818 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="sg-core" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984850 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="ceilometer-notification-agent" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.984865 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" containerName="proxy-httpd" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.985619 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.988062 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 06 09:02:21 crc kubenswrapper[4989]: I1006 09:02:21.993439 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ded3-account-create-2p5m9"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.047580 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hkcz\" (UniqueName: \"kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz\") pod \"nova-api-ded3-account-create-2p5m9\" (UID: \"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c\") " pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.133787 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe","Type":"ContainerDied","Data":"daf5b8c9bba19f7a8e55d18507fc3f32a633710d7290a1c92ce4a8d2511e7802"} Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.133842 4989 scope.go:117] "RemoveContainer" containerID="6b63e8b67a9d19855d3c01aa43cd7cfb864abb3738572d14b99446cf360b4da7" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.133807 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.137493 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ef8cd741-ab58-4662-872c-8955c47ef9ca","Type":"ContainerDied","Data":"94377efccbefe338de524e9a0f206ca4ee08411cad63b480db8cc334e837a262"} Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.137541 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.150425 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hkcz\" (UniqueName: \"kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz\") pod \"nova-api-ded3-account-create-2p5m9\" (UID: \"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c\") " pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.161217 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.171380 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.175678 4989 scope.go:117] "RemoveContainer" containerID="7d67e6989f2726f5fff92491c13e5646677be4b188c274a06e93bde0e31ed8f4" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.181762 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ef11-account-create-lp9dg"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.182955 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.191969 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef11-account-create-lp9dg"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.202139 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.203833 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.206909 4989 scope.go:117] "RemoveContainer" containerID="3061bb1ec9fc4bfdf9250936563634f747e72909374e0d922e1b15d16d70592b" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.209804 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hkcz\" (UniqueName: \"kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz\") pod \"nova-api-ded3-account-create-2p5m9\" (UID: \"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c\") " pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.222571 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.241734 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.243628 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.257218 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cssbd\" (UniqueName: \"kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd\") pod \"nova-cell0-ef11-account-create-lp9dg\" (UID: \"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd\") " pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.257385 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.257700 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.281701 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.302339 4989 scope.go:117] "RemoveContainer" containerID="96fed2b62148d2ee8435c799fb4d5985da17893469900df4534055b6130b75c1" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.307026 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.327967 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.334021 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.336603 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.336950 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.337030 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.337734 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.358924 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.358994 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359078 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359107 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359193 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359246 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359294 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cssbd\" (UniqueName: \"kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd\") pod \"nova-cell0-ef11-account-create-lp9dg\" (UID: \"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd\") " pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.359333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk6v9\" (UniqueName: \"kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.368443 4989 scope.go:117] "RemoveContainer" containerID="40f66117b81ed15de75788b89b54374fee5f59e5999963b377716605187461ed" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.394796 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cssbd\" (UniqueName: \"kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd\") pod \"nova-cell0-ef11-account-create-lp9dg\" (UID: \"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd\") " pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.401580 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4c6e-account-create-zhzb5"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.402744 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.405937 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.434417 4989 scope.go:117] "RemoveContainer" containerID="1280c170a12709fa4a2cda4733cc078a2937e3460f0d2fb63d5c9b83282fee7f" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.454368 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4c6e-account-create-zhzb5"] Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.459016 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460186 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk6v9\" (UniqueName: \"kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460210 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460237 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460263 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460279 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460298 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460331 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460349 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460375 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460392 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460413 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460467 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kfc2\" (UniqueName: \"kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2\") pod \"nova-cell1-4c6e-account-create-zhzb5\" (UID: \"8f6adc1f-6029-46fe-939c-8633e05cb14b\") " pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460515 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460532 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmmhz\" (UniqueName: \"kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460550 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.460597 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.462465 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.463260 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.464400 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.465759 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.477619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.477993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.481040 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk6v9\" (UniqueName: \"kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.481212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.519969 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562626 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562804 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562837 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.562949 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kfc2\" (UniqueName: \"kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2\") pod \"nova-cell1-4c6e-account-create-zhzb5\" (UID: \"8f6adc1f-6029-46fe-939c-8633e05cb14b\") " pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.563056 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmmhz\" (UniqueName: \"kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.563089 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.563140 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.566469 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.566751 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.566961 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.567189 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.569031 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.570700 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.570776 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.578471 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.583515 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kfc2\" (UniqueName: \"kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2\") pod \"nova-cell1-4c6e-account-create-zhzb5\" (UID: \"8f6adc1f-6029-46fe-939c-8633e05cb14b\") " pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.591926 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmmhz\" (UniqueName: \"kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz\") pod \"ceilometer-0\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.730136 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.775191 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.792168 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:22 crc kubenswrapper[4989]: W1006 09:02:22.863317 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5c7c63e_55e8_47f0_bf26_f5b83ea54b4c.slice/crio-1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89 WatchSource:0}: Error finding container 1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89: Status 404 returned error can't find the container with id 1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89 Oct 06 09:02:22 crc kubenswrapper[4989]: I1006 09:02:22.879754 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ded3-account-create-2p5m9"] Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.070082 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ef11-account-create-lp9dg"] Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.185313 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef11-account-create-lp9dg" event={"ID":"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd","Type":"ContainerStarted","Data":"68478eb8b4ff20f0e085935d0d264bab9ce15fb747e66552a514e189aa64cd40"} Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.192638 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ded3-account-create-2p5m9" event={"ID":"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c","Type":"ContainerStarted","Data":"42b5b43509a943241d2558a5a7d5a648587db28f3877b2329b996d7e657a1eea"} Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.193543 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ded3-account-create-2p5m9" event={"ID":"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c","Type":"ContainerStarted","Data":"1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89"} Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.208342 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-ded3-account-create-2p5m9" podStartSLOduration=2.208319116 podStartE2EDuration="2.208319116s" podCreationTimestamp="2025-10-06 09:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:23.206218365 +0000 UTC m=+1393.996243945" watchObservedRunningTime="2025-10-06 09:02:23.208319116 +0000 UTC m=+1393.998344696" Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.351900 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.373672 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:23 crc kubenswrapper[4989]: W1006 09:02:23.377436 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75b2344f_bdb4_4854_a732_6069c50f41f2.slice/crio-d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee WatchSource:0}: Error finding container d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee: Status 404 returned error can't find the container with id d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.450448 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4c6e-account-create-zhzb5"] Oct 06 09:02:23 crc kubenswrapper[4989]: W1006 09:02:23.454950 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f6adc1f_6029_46fe_939c_8633e05cb14b.slice/crio-4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695 WatchSource:0}: Error finding container 4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695: Status 404 returned error can't find the container with id 4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695 Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.516540 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.949198 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe" path="/var/lib/kubelet/pods/1f433a70-1070-4c6c-ad1a-fa6bd5fb78fe/volumes" Oct 06 09:02:23 crc kubenswrapper[4989]: I1006 09:02:23.950676 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef8cd741-ab58-4662-872c-8955c47ef9ca" path="/var/lib/kubelet/pods/ef8cd741-ab58-4662-872c-8955c47ef9ca/volumes" Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.251776 4989 generic.go:334] "Generic (PLEG): container finished" podID="093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" containerID="b5185fa6bcba77dca8b60346a7556b03433f6545c232af16b93dcb0b73118fd4" exitCode=0 Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.251960 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef11-account-create-lp9dg" event={"ID":"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd","Type":"ContainerDied","Data":"b5185fa6bcba77dca8b60346a7556b03433f6545c232af16b93dcb0b73118fd4"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.254484 4989 generic.go:334] "Generic (PLEG): container finished" podID="e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" containerID="42b5b43509a943241d2558a5a7d5a648587db28f3877b2329b996d7e657a1eea" exitCode=0 Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.254611 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ded3-account-create-2p5m9" event={"ID":"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c","Type":"ContainerDied","Data":"42b5b43509a943241d2558a5a7d5a648587db28f3877b2329b996d7e657a1eea"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.257160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerStarted","Data":"3814cda71279e66c461650e21aeab490084205cdc04eb79209f77094180bab6e"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.257203 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerStarted","Data":"d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.260310 4989 generic.go:334] "Generic (PLEG): container finished" podID="8f6adc1f-6029-46fe-939c-8633e05cb14b" containerID="bc352a1f9f91a5b814eb79f29c22ab04be980b414e6d6507bfcec57d13be32b1" exitCode=0 Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.260353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" event={"ID":"8f6adc1f-6029-46fe-939c-8633e05cb14b","Type":"ContainerDied","Data":"bc352a1f9f91a5b814eb79f29c22ab04be980b414e6d6507bfcec57d13be32b1"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.260370 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" event={"ID":"8f6adc1f-6029-46fe-939c-8633e05cb14b","Type":"ContainerStarted","Data":"4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.266909 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerStarted","Data":"371b119db5c7bdd79dd95a2f018f52c39640a38650875d86dba0c44d26a5cf87"} Oct 06 09:02:24 crc kubenswrapper[4989]: I1006 09:02:24.266940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerStarted","Data":"a071fe7b3fe26a8edada0088ec027da7c763e2c910b425ad8ccf666702f89caa"} Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.277723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerStarted","Data":"a7ea9a251ec17f10791afe94beed6956bf60120b934e4c1b26a49ceedb4b0502"} Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.280405 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerStarted","Data":"ade5f06f53063a0b3ae8f9d532b65a48a6cd63581297c71cb12865f82e8d69a2"} Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.301696 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.301679298 podStartE2EDuration="3.301679298s" podCreationTimestamp="2025-10-06 09:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:25.297716644 +0000 UTC m=+1396.087742244" watchObservedRunningTime="2025-10-06 09:02:25.301679298 +0000 UTC m=+1396.091704868" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.656637 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.704264 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.752111 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.840747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cssbd\" (UniqueName: \"kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd\") pod \"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd\" (UID: \"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd\") " Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.840973 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kfc2\" (UniqueName: \"kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2\") pod \"8f6adc1f-6029-46fe-939c-8633e05cb14b\" (UID: \"8f6adc1f-6029-46fe-939c-8633e05cb14b\") " Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.847894 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2" (OuterVolumeSpecName: "kube-api-access-8kfc2") pod "8f6adc1f-6029-46fe-939c-8633e05cb14b" (UID: "8f6adc1f-6029-46fe-939c-8633e05cb14b"). InnerVolumeSpecName "kube-api-access-8kfc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.847934 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd" (OuterVolumeSpecName: "kube-api-access-cssbd") pod "093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" (UID: "093ba9d5-8d9d-4e0d-837b-21575e2ad9dd"). InnerVolumeSpecName "kube-api-access-cssbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.943109 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hkcz\" (UniqueName: \"kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz\") pod \"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c\" (UID: \"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c\") " Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.944689 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kfc2\" (UniqueName: \"kubernetes.io/projected/8f6adc1f-6029-46fe-939c-8633e05cb14b-kube-api-access-8kfc2\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.944707 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cssbd\" (UniqueName: \"kubernetes.io/projected/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd-kube-api-access-cssbd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:25 crc kubenswrapper[4989]: I1006 09:02:25.948896 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz" (OuterVolumeSpecName: "kube-api-access-4hkcz") pod "e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" (UID: "e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c"). InnerVolumeSpecName "kube-api-access-4hkcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.049556 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hkcz\" (UniqueName: \"kubernetes.io/projected/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c-kube-api-access-4hkcz\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.291049 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ded3-account-create-2p5m9" event={"ID":"e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c","Type":"ContainerDied","Data":"1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89"} Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.291097 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a51e7160f6136d7f79d48f287ca32f5c663e66687c1930071e625c8a34b5a89" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.291065 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ded3-account-create-2p5m9" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.292490 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" event={"ID":"8f6adc1f-6029-46fe-939c-8633e05cb14b","Type":"ContainerDied","Data":"4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695"} Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.292522 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b158b266ab3466f73395c35068007b7507470b828456749dd2b038d763a2695" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.292801 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4c6e-account-create-zhzb5" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.294718 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerStarted","Data":"1119310c350c80492398124a9f6e92e962cedfc2deda815f809c15ae3a6861da"} Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.296715 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ef11-account-create-lp9dg" event={"ID":"093ba9d5-8d9d-4e0d-837b-21575e2ad9dd","Type":"ContainerDied","Data":"68478eb8b4ff20f0e085935d0d264bab9ce15fb747e66552a514e189aa64cd40"} Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.296766 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68478eb8b4ff20f0e085935d0d264bab9ce15fb747e66552a514e189aa64cd40" Oct 06 09:02:26 crc kubenswrapper[4989]: I1006 09:02:26.296741 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ef11-account-create-lp9dg" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.345273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerStarted","Data":"7e592533c8536d85c7490d9369c9e8093165f702f677da64db52736063dbe202"} Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.345870 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-central-agent" containerID="cri-o://371b119db5c7bdd79dd95a2f018f52c39640a38650875d86dba0c44d26a5cf87" gracePeriod=30 Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.346034 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.346519 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="proxy-httpd" containerID="cri-o://7e592533c8536d85c7490d9369c9e8093165f702f677da64db52736063dbe202" gracePeriod=30 Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.346600 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-notification-agent" containerID="cri-o://ade5f06f53063a0b3ae8f9d532b65a48a6cd63581297c71cb12865f82e8d69a2" gracePeriod=30 Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.346708 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="sg-core" containerID="cri-o://1119310c350c80492398124a9f6e92e962cedfc2deda815f809c15ae3a6861da" gracePeriod=30 Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.355597 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5lp9k"] Oct 06 09:02:27 crc kubenswrapper[4989]: E1006 09:02:27.362240 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f6adc1f-6029-46fe-939c-8633e05cb14b" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362277 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f6adc1f-6029-46fe-939c-8633e05cb14b" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: E1006 09:02:27.362299 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362305 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: E1006 09:02:27.362325 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362331 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362603 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362621 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.362635 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f6adc1f-6029-46fe-939c-8633e05cb14b" containerName="mariadb-account-create" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.363254 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.365502 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.365673 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nxpft" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.366025 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5lp9k"] Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.366941 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.389394 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.996213308 podStartE2EDuration="5.389372868s" podCreationTimestamp="2025-10-06 09:02:22 +0000 UTC" firstStartedPulling="2025-10-06 09:02:23.38391163 +0000 UTC m=+1394.173937210" lastFinishedPulling="2025-10-06 09:02:26.77707119 +0000 UTC m=+1397.567096770" observedRunningTime="2025-10-06 09:02:27.388205204 +0000 UTC m=+1398.178230784" watchObservedRunningTime="2025-10-06 09:02:27.389372868 +0000 UTC m=+1398.179398448" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.474961 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.475051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.475402 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.475643 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.577889 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.578293 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.578319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.578448 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.583136 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.583281 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.583960 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.595716 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh\") pod \"nova-cell0-conductor-db-sync-5lp9k\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:27 crc kubenswrapper[4989]: I1006 09:02:27.679068 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.139969 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5lp9k"] Oct 06 09:02:28 crc kubenswrapper[4989]: W1006 09:02:28.142689 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a87fd5d_ae89_4d9f_aa82_1ef0380d9f0e.slice/crio-7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9 WatchSource:0}: Error finding container 7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9: Status 404 returned error can't find the container with id 7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9 Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.356252 4989 generic.go:334] "Generic (PLEG): container finished" podID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerID="7e592533c8536d85c7490d9369c9e8093165f702f677da64db52736063dbe202" exitCode=0 Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.357403 4989 generic.go:334] "Generic (PLEG): container finished" podID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerID="1119310c350c80492398124a9f6e92e962cedfc2deda815f809c15ae3a6861da" exitCode=2 Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.357537 4989 generic.go:334] "Generic (PLEG): container finished" podID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerID="ade5f06f53063a0b3ae8f9d532b65a48a6cd63581297c71cb12865f82e8d69a2" exitCode=0 Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.357378 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerDied","Data":"7e592533c8536d85c7490d9369c9e8093165f702f677da64db52736063dbe202"} Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.357827 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerDied","Data":"1119310c350c80492398124a9f6e92e962cedfc2deda815f809c15ae3a6861da"} Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.357913 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerDied","Data":"ade5f06f53063a0b3ae8f9d532b65a48a6cd63581297c71cb12865f82e8d69a2"} Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.358954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" event={"ID":"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e","Type":"ContainerStarted","Data":"7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9"} Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.479254 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.479556 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.509592 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 09:02:28 crc kubenswrapper[4989]: I1006 09:02:28.518715 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 09:02:29 crc kubenswrapper[4989]: I1006 09:02:29.367222 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 09:02:29 crc kubenswrapper[4989]: I1006 09:02:29.367559 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 09:02:31 crc kubenswrapper[4989]: I1006 09:02:31.334449 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 09:02:31 crc kubenswrapper[4989]: I1006 09:02:31.384605 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 09:02:31 crc kubenswrapper[4989]: I1006 09:02:31.459136 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.394993 4989 generic.go:334] "Generic (PLEG): container finished" podID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerID="371b119db5c7bdd79dd95a2f018f52c39640a38650875d86dba0c44d26a5cf87" exitCode=0 Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.395073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerDied","Data":"371b119db5c7bdd79dd95a2f018f52c39640a38650875d86dba0c44d26a5cf87"} Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.730765 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.730816 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.772491 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:32 crc kubenswrapper[4989]: I1006 09:02:32.777917 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:33 crc kubenswrapper[4989]: I1006 09:02:33.405270 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:33 crc kubenswrapper[4989]: I1006 09:02:33.405609 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:34 crc kubenswrapper[4989]: I1006 09:02:34.895502 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.018770 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.019903 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020015 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020041 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmmhz\" (UniqueName: \"kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020127 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020169 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020190 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020265 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle\") pod \"1a256989-88f3-4e04-8927-ead6e4939cc6\" (UID: \"1a256989-88f3-4e04-8927-ead6e4939cc6\") " Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.020973 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.021268 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.024747 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz" (OuterVolumeSpecName: "kube-api-access-pmmhz") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "kube-api-access-pmmhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.027068 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts" (OuterVolumeSpecName: "scripts") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.053084 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.098863 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122239 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122276 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a256989-88f3-4e04-8927-ead6e4939cc6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122286 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122300 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122312 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.122325 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmmhz\" (UniqueName: \"kubernetes.io/projected/1a256989-88f3-4e04-8927-ead6e4939cc6-kube-api-access-pmmhz\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.123635 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.129935 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data" (OuterVolumeSpecName: "config-data") pod "1a256989-88f3-4e04-8927-ead6e4939cc6" (UID: "1a256989-88f3-4e04-8927-ead6e4939cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.223994 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.224039 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a256989-88f3-4e04-8927-ead6e4939cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.420944 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.431044 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.441939 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a256989-88f3-4e04-8927-ead6e4939cc6","Type":"ContainerDied","Data":"a071fe7b3fe26a8edada0088ec027da7c763e2c910b425ad8ccf666702f89caa"} Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.442019 4989 scope.go:117] "RemoveContainer" containerID="7e592533c8536d85c7490d9369c9e8093165f702f677da64db52736063dbe202" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.442151 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.444021 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" event={"ID":"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e","Type":"ContainerStarted","Data":"0ba87e5984d61addcfc335a8689464e448fcc9f8d299b926cd462ad4453f562e"} Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.470525 4989 scope.go:117] "RemoveContainer" containerID="1119310c350c80492398124a9f6e92e962cedfc2deda815f809c15ae3a6861da" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.496097 4989 scope.go:117] "RemoveContainer" containerID="ade5f06f53063a0b3ae8f9d532b65a48a6cd63581297c71cb12865f82e8d69a2" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.519904 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" podStartSLOduration=2.104447221 podStartE2EDuration="8.519878374s" podCreationTimestamp="2025-10-06 09:02:27 +0000 UTC" firstStartedPulling="2025-10-06 09:02:28.144710222 +0000 UTC m=+1398.934735802" lastFinishedPulling="2025-10-06 09:02:34.560141375 +0000 UTC m=+1405.350166955" observedRunningTime="2025-10-06 09:02:35.49338288 +0000 UTC m=+1406.283408460" watchObservedRunningTime="2025-10-06 09:02:35.519878374 +0000 UTC m=+1406.309903954" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.540487 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.550703 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.562680 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:35 crc kubenswrapper[4989]: E1006 09:02:35.563101 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="sg-core" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563125 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="sg-core" Oct 06 09:02:35 crc kubenswrapper[4989]: E1006 09:02:35.563144 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-notification-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563154 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-notification-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: E1006 09:02:35.563173 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-central-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563182 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-central-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: E1006 09:02:35.563215 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="proxy-httpd" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563223 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="proxy-httpd" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563429 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="proxy-httpd" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563444 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="sg-core" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563454 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-central-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.563462 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" containerName="ceilometer-notification-agent" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.581040 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.581176 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.590532 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.590861 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.591057 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.599051 4989 scope.go:117] "RemoveContainer" containerID="371b119db5c7bdd79dd95a2f018f52c39640a38650875d86dba0c44d26a5cf87" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.631957 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632017 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632039 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632085 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632107 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632134 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt8kk\" (UniqueName: \"kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.632183 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733350 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733397 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733436 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt8kk\" (UniqueName: \"kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733589 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733622 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.733641 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.734523 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.734947 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.737780 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.738076 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.738890 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.740675 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.743590 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.756366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt8kk\" (UniqueName: \"kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk\") pod \"ceilometer-0\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.914621 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:02:35 crc kubenswrapper[4989]: I1006 09:02:35.962438 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a256989-88f3-4e04-8927-ead6e4939cc6" path="/var/lib/kubelet/pods/1a256989-88f3-4e04-8927-ead6e4939cc6/volumes" Oct 06 09:02:36 crc kubenswrapper[4989]: W1006 09:02:36.415857 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c60c27d_e549_4ced_a2eb_6e400cb110cc.slice/crio-9c4a64bb2ee45c6cf2ee200ac852959d7502554f9a2c1442f4536087ac8d14fe WatchSource:0}: Error finding container 9c4a64bb2ee45c6cf2ee200ac852959d7502554f9a2c1442f4536087ac8d14fe: Status 404 returned error can't find the container with id 9c4a64bb2ee45c6cf2ee200ac852959d7502554f9a2c1442f4536087ac8d14fe Oct 06 09:02:36 crc kubenswrapper[4989]: I1006 09:02:36.416796 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:02:36 crc kubenswrapper[4989]: I1006 09:02:36.458408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerStarted","Data":"9c4a64bb2ee45c6cf2ee200ac852959d7502554f9a2c1442f4536087ac8d14fe"} Oct 06 09:02:37 crc kubenswrapper[4989]: I1006 09:02:37.474248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerStarted","Data":"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d"} Oct 06 09:02:38 crc kubenswrapper[4989]: I1006 09:02:38.483980 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerStarted","Data":"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208"} Oct 06 09:02:38 crc kubenswrapper[4989]: I1006 09:02:38.484348 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerStarted","Data":"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30"} Oct 06 09:02:40 crc kubenswrapper[4989]: I1006 09:02:40.509327 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerStarted","Data":"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07"} Oct 06 09:02:40 crc kubenswrapper[4989]: I1006 09:02:40.510098 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 09:02:40 crc kubenswrapper[4989]: I1006 09:02:40.559958 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.454196451 podStartE2EDuration="5.559930331s" podCreationTimestamp="2025-10-06 09:02:35 +0000 UTC" firstStartedPulling="2025-10-06 09:02:36.417743569 +0000 UTC m=+1407.207769139" lastFinishedPulling="2025-10-06 09:02:39.523477439 +0000 UTC m=+1410.313503019" observedRunningTime="2025-10-06 09:02:40.537395131 +0000 UTC m=+1411.327420711" watchObservedRunningTime="2025-10-06 09:02:40.559930331 +0000 UTC m=+1411.349955911" Oct 06 09:02:46 crc kubenswrapper[4989]: I1006 09:02:46.573735 4989 generic.go:334] "Generic (PLEG): container finished" podID="9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" containerID="0ba87e5984d61addcfc335a8689464e448fcc9f8d299b926cd462ad4453f562e" exitCode=0 Oct 06 09:02:46 crc kubenswrapper[4989]: I1006 09:02:46.573873 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" event={"ID":"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e","Type":"ContainerDied","Data":"0ba87e5984d61addcfc335a8689464e448fcc9f8d299b926cd462ad4453f562e"} Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.943187 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.966210 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle\") pod \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.966328 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts\") pod \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.966397 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh\") pod \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.966457 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data\") pod \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\" (UID: \"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e\") " Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.972824 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts" (OuterVolumeSpecName: "scripts") pod "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" (UID: "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.973748 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh" (OuterVolumeSpecName: "kube-api-access-fn8fh") pod "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" (UID: "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e"). InnerVolumeSpecName "kube-api-access-fn8fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:02:47 crc kubenswrapper[4989]: I1006 09:02:47.999742 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data" (OuterVolumeSpecName: "config-data") pod "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" (UID: "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.010517 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" (UID: "9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.068559 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.068611 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn8fh\" (UniqueName: \"kubernetes.io/projected/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-kube-api-access-fn8fh\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.068625 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.068638 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.598730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" event={"ID":"9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e","Type":"ContainerDied","Data":"7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9"} Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.599015 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bd69003c00415e6918960e27b780e80c9f554b4d11d7f129ecf9dffdfb7cee9" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.598782 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5lp9k" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.694107 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:02:48 crc kubenswrapper[4989]: E1006 09:02:48.694585 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" containerName="nova-cell0-conductor-db-sync" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.694606 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" containerName="nova-cell0-conductor-db-sync" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.694856 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" containerName="nova-cell0-conductor-db-sync" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.695484 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.697597 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nxpft" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.701582 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.707217 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.783724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.783823 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8w6w\" (UniqueName: \"kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.783972 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.885516 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.885619 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8w6w\" (UniqueName: \"kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.885750 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.891869 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.893017 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:48 crc kubenswrapper[4989]: I1006 09:02:48.902161 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8w6w\" (UniqueName: \"kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w\") pod \"nova-cell0-conductor-0\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:49 crc kubenswrapper[4989]: I1006 09:02:49.015931 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:49 crc kubenswrapper[4989]: I1006 09:02:49.497119 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:02:49 crc kubenswrapper[4989]: I1006 09:02:49.610451 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5a17496d-00da-47b7-a178-5274afa8ac00","Type":"ContainerStarted","Data":"1911e063ec7efbf835478a5f0ace21b2f1ac61866b1fec81f151df77f0beb741"} Oct 06 09:02:50 crc kubenswrapper[4989]: I1006 09:02:50.620339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5a17496d-00da-47b7-a178-5274afa8ac00","Type":"ContainerStarted","Data":"275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb"} Oct 06 09:02:50 crc kubenswrapper[4989]: I1006 09:02:50.621729 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:50 crc kubenswrapper[4989]: I1006 09:02:50.642301 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.642280759 podStartE2EDuration="2.642280759s" podCreationTimestamp="2025-10-06 09:02:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:50.63470414 +0000 UTC m=+1421.424729730" watchObservedRunningTime="2025-10-06 09:02:50.642280759 +0000 UTC m=+1421.432306329" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.057339 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.536060 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-mwhkn"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.538969 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.542568 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.546096 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.561333 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mwhkn"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.683445 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcgg8\" (UniqueName: \"kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.683840 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.684089 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.684250 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.701814 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.704154 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.714193 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.719258 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.767324 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.768946 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.776501 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.786097 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.787952 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.788135 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.788354 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcgg8\" (UniqueName: \"kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.791224 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.793546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.793549 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.801418 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.854769 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcgg8\" (UniqueName: \"kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8\") pod \"nova-cell0-cell-mapping-mwhkn\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.875137 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.890698 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.890759 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.890887 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.890951 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.890993 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.891100 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dwqq\" (UniqueName: \"kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.891161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngcf\" (UniqueName: \"kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.897753 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.899825 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.905862 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 09:02:54 crc kubenswrapper[4989]: I1006 09:02:54.929537 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.998320 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.998406 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.998448 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.998936 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dwqq\" (UniqueName: \"kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.999004 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngcf\" (UniqueName: \"kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.999092 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:54.999121 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.008386 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.008531 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.010013 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.018337 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.018902 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.018919 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.018991 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.023669 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.030476 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.035214 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.036758 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.041440 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dwqq\" (UniqueName: \"kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq\") pod \"nova-scheduler-0\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.049402 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.052427 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngcf\" (UniqueName: \"kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf\") pod \"nova-api-0\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.100573 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.100735 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.100831 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crnq\" (UniqueName: \"kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.100893 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.201561 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202478 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crnq\" (UniqueName: \"kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202537 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202564 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202583 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202615 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkmll\" (UniqueName: \"kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202663 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202701 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202727 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202760 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202787 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202827 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp48m\" (UniqueName: \"kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.202860 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.204108 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.210059 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.213807 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.234559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crnq\" (UniqueName: \"kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq\") pod \"nova-metadata-0\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.277328 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304593 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304686 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304722 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304740 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp48m\" (UniqueName: \"kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304768 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304810 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304834 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304863 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkmll\" (UniqueName: \"kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.304914 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.306762 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.309300 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.312870 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.313827 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.313858 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.313996 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.321896 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.324824 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkmll\" (UniqueName: \"kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.328851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp48m\" (UniqueName: \"kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m\") pod \"dnsmasq-dns-845d6d6f59-chph9\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.336977 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.437981 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.442472 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.567209 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mwhkn"] Oct 06 09:02:55 crc kubenswrapper[4989]: W1006 09:02:55.572271 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb416f76a_8dcd_411a_a0a0_31f7becea55e.slice/crio-e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062 WatchSource:0}: Error finding container e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062: Status 404 returned error can't find the container with id e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062 Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.700268 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mwhkn" event={"ID":"b416f76a-8dcd-411a-a0a0-31f7becea55e","Type":"ContainerStarted","Data":"e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062"} Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.730721 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-r7w8z"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.732062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.735939 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.738827 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.742869 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-r7w8z"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.779117 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:02:55 crc kubenswrapper[4989]: W1006 09:02:55.880327 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd61f7286_65ea_4ed6_8d18_b4edf8afe373.slice/crio-640c973f1b77cd291520ddea2cfd31a0dee7ef1793833d7e6f2b839ce287cfc2 WatchSource:0}: Error finding container 640c973f1b77cd291520ddea2cfd31a0dee7ef1793833d7e6f2b839ce287cfc2: Status 404 returned error can't find the container with id 640c973f1b77cd291520ddea2cfd31a0dee7ef1793833d7e6f2b839ce287cfc2 Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.889508 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.934618 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.934713 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.934775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.934882 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ftv\" (UniqueName: \"kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:55 crc kubenswrapper[4989]: I1006 09:02:55.964565 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.036711 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ftv\" (UniqueName: \"kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.036909 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.036926 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.036963 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.048694 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.048758 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.049197 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.051212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.061260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ftv\" (UniqueName: \"kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv\") pod \"nova-cell1-conductor-db-sync-r7w8z\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.086130 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.133060 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.717188 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-r7w8z"] Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.723446 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4386782-615c-4aa0-960f-83c312cbb724","Type":"ContainerStarted","Data":"baef450d5c104aa1305dcffacc6be9bdbf6a983c93211f900f7dd8f78ae03c1d"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.725212 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerStarted","Data":"640c973f1b77cd291520ddea2cfd31a0dee7ef1793833d7e6f2b839ce287cfc2"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.726743 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"377dacd9-0ff0-4fab-8cbd-e28f205efb0d","Type":"ContainerStarted","Data":"781f6c70c9481c574bb4fa7d81df54dd605c61718e54f5d1ae2c126620b0d0ac"} Oct 06 09:02:56 crc kubenswrapper[4989]: W1006 09:02:56.729380 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aec9ce1_0a3f_4788_a367_4fa2f1554d44.slice/crio-33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243 WatchSource:0}: Error finding container 33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243: Status 404 returned error can't find the container with id 33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243 Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.729814 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mwhkn" event={"ID":"b416f76a-8dcd-411a-a0a0-31f7becea55e","Type":"ContainerStarted","Data":"58c5d959d29623aea9a09ac52e22bed81527dfc5edac480b7610c6595fb84cfd"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.732604 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerStarted","Data":"ab7a9f16c03f909d5a07a3b955c22bb5d3e28beff36c00d32d86b61f1e9ee8fd"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.736483 4989 generic.go:334] "Generic (PLEG): container finished" podID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerID="0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a" exitCode=0 Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.736526 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" event={"ID":"8459d522-eb54-4580-b3ca-39e54c86a3c5","Type":"ContainerDied","Data":"0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.736555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" event={"ID":"8459d522-eb54-4580-b3ca-39e54c86a3c5","Type":"ContainerStarted","Data":"a541047321fd3b6cc60b201f97e426d45c82cb179cff3f28a78ef86d776e9134"} Oct 06 09:02:56 crc kubenswrapper[4989]: I1006 09:02:56.759933 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-mwhkn" podStartSLOduration=2.759857791 podStartE2EDuration="2.759857791s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:56.74731032 +0000 UTC m=+1427.537335900" watchObservedRunningTime="2025-10-06 09:02:56.759857791 +0000 UTC m=+1427.549883371" Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.754526 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" event={"ID":"8459d522-eb54-4580-b3ca-39e54c86a3c5","Type":"ContainerStarted","Data":"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba"} Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.755145 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.759549 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" event={"ID":"8aec9ce1-0a3f-4788-a367-4fa2f1554d44","Type":"ContainerStarted","Data":"16a49375a3c8f1a4186ec2a1c51cfb2a9d9d97c9799db9d7fe6d33abe94affef"} Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.759594 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" event={"ID":"8aec9ce1-0a3f-4788-a367-4fa2f1554d44","Type":"ContainerStarted","Data":"33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243"} Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.778381 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" podStartSLOduration=3.7783578159999998 podStartE2EDuration="3.778357816s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:57.772793735 +0000 UTC m=+1428.562819335" watchObservedRunningTime="2025-10-06 09:02:57.778357816 +0000 UTC m=+1428.568383396" Oct 06 09:02:57 crc kubenswrapper[4989]: I1006 09:02:57.794075 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" podStartSLOduration=2.794058558 podStartE2EDuration="2.794058558s" podCreationTimestamp="2025-10-06 09:02:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:02:57.790975449 +0000 UTC m=+1428.581001039" watchObservedRunningTime="2025-10-06 09:02:57.794058558 +0000 UTC m=+1428.584084138" Oct 06 09:02:58 crc kubenswrapper[4989]: I1006 09:02:58.749984 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:02:58 crc kubenswrapper[4989]: I1006 09:02:58.761422 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.781493 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4386782-615c-4aa0-960f-83c312cbb724","Type":"ContainerStarted","Data":"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.793763 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerStarted","Data":"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.793916 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerStarted","Data":"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.794327 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-log" containerID="cri-o://1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" gracePeriod=30 Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.794330 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-metadata" containerID="cri-o://4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" gracePeriod=30 Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.822863 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"377dacd9-0ff0-4fab-8cbd-e28f205efb0d","Type":"ContainerStarted","Data":"43c2e169810eb04bada5f35eaaf81350cd70184b618b76a271f06ac3f1d73440"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.823018 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://43c2e169810eb04bada5f35eaaf81350cd70184b618b76a271f06ac3f1d73440" gracePeriod=30 Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.829324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerStarted","Data":"ebb72deedd186195f6b578c0c2fa80fae603ae9607847a578e8bb536f1883b67"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.829361 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerStarted","Data":"26863d5b0d9b2d53be9b4e56d60c8e66f73ac31b5497f62f6b98628cde8cac5d"} Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.832347 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.915984043 podStartE2EDuration="5.832328562s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="2025-10-06 09:02:55.781213947 +0000 UTC m=+1426.571239537" lastFinishedPulling="2025-10-06 09:02:58.697558486 +0000 UTC m=+1429.487584056" observedRunningTime="2025-10-06 09:02:59.801369039 +0000 UTC m=+1430.591394629" watchObservedRunningTime="2025-10-06 09:02:59.832328562 +0000 UTC m=+1430.622354142" Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.842275 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.008773019 podStartE2EDuration="5.842234288s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="2025-10-06 09:02:55.882881719 +0000 UTC m=+1426.672907299" lastFinishedPulling="2025-10-06 09:02:58.716342988 +0000 UTC m=+1429.506368568" observedRunningTime="2025-10-06 09:02:59.82221069 +0000 UTC m=+1430.612236290" watchObservedRunningTime="2025-10-06 09:02:59.842234288 +0000 UTC m=+1430.632259868" Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.859265 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.212588228 podStartE2EDuration="5.859250319s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="2025-10-06 09:02:56.050220575 +0000 UTC m=+1426.840246155" lastFinishedPulling="2025-10-06 09:02:58.696882656 +0000 UTC m=+1429.486908246" observedRunningTime="2025-10-06 09:02:59.844023419 +0000 UTC m=+1430.634048999" watchObservedRunningTime="2025-10-06 09:02:59.859250319 +0000 UTC m=+1430.649275899" Oct 06 09:02:59 crc kubenswrapper[4989]: I1006 09:02:59.876193 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.137266526 podStartE2EDuration="5.876174367s" podCreationTimestamp="2025-10-06 09:02:54 +0000 UTC" firstStartedPulling="2025-10-06 09:02:55.955337329 +0000 UTC m=+1426.745362909" lastFinishedPulling="2025-10-06 09:02:58.69424517 +0000 UTC m=+1429.484270750" observedRunningTime="2025-10-06 09:02:59.864701226 +0000 UTC m=+1430.654726816" watchObservedRunningTime="2025-10-06 09:02:59.876174367 +0000 UTC m=+1430.666199947" Oct 06 09:02:59 crc kubenswrapper[4989]: E1006 09:02:59.885197 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd61f7286_65ea_4ed6_8d18_b4edf8afe373.slice/crio-conmon-1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667.scope\": RecentStats: unable to find data in memory cache]" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.202951 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.278678 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.278736 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.420776 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.438255 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528211 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crnq\" (UniqueName: \"kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq\") pod \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528278 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle\") pod \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528313 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs\") pod \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528361 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data\") pod \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\" (UID: \"d61f7286-65ea-4ed6-8d18-b4edf8afe373\") " Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528633 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs" (OuterVolumeSpecName: "logs") pod "d61f7286-65ea-4ed6-8d18-b4edf8afe373" (UID: "d61f7286-65ea-4ed6-8d18-b4edf8afe373"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.528791 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d61f7286-65ea-4ed6-8d18-b4edf8afe373-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.533835 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq" (OuterVolumeSpecName: "kube-api-access-2crnq") pod "d61f7286-65ea-4ed6-8d18-b4edf8afe373" (UID: "d61f7286-65ea-4ed6-8d18-b4edf8afe373"). InnerVolumeSpecName "kube-api-access-2crnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.560794 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d61f7286-65ea-4ed6-8d18-b4edf8afe373" (UID: "d61f7286-65ea-4ed6-8d18-b4edf8afe373"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.567644 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data" (OuterVolumeSpecName: "config-data") pod "d61f7286-65ea-4ed6-8d18-b4edf8afe373" (UID: "d61f7286-65ea-4ed6-8d18-b4edf8afe373"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.630725 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crnq\" (UniqueName: \"kubernetes.io/projected/d61f7286-65ea-4ed6-8d18-b4edf8afe373-kube-api-access-2crnq\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.630773 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.630785 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d61f7286-65ea-4ed6-8d18-b4edf8afe373-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842052 4989 generic.go:334] "Generic (PLEG): container finished" podID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerID="4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" exitCode=0 Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842097 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842127 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerDied","Data":"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7"} Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842170 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerDied","Data":"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667"} Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842198 4989 scope.go:117] "RemoveContainer" containerID="4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842104 4989 generic.go:334] "Generic (PLEG): container finished" podID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerID="1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" exitCode=143 Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.842426 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d61f7286-65ea-4ed6-8d18-b4edf8afe373","Type":"ContainerDied","Data":"640c973f1b77cd291520ddea2cfd31a0dee7ef1793833d7e6f2b839ce287cfc2"} Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.884376 4989 scope.go:117] "RemoveContainer" containerID="1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.892460 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.911077 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.924507 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:00 crc kubenswrapper[4989]: E1006 09:03:00.925714 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-log" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.925739 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-log" Oct 06 09:03:00 crc kubenswrapper[4989]: E1006 09:03:00.925754 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-metadata" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.925762 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-metadata" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.925987 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-log" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.926016 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" containerName="nova-metadata-metadata" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.927980 4989 scope.go:117] "RemoveContainer" containerID="4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" Oct 06 09:03:00 crc kubenswrapper[4989]: E1006 09:03:00.928510 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7\": container with ID starting with 4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7 not found: ID does not exist" containerID="4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.928572 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7"} err="failed to get container status \"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7\": rpc error: code = NotFound desc = could not find container \"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7\": container with ID starting with 4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7 not found: ID does not exist" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.928608 4989 scope.go:117] "RemoveContainer" containerID="1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.929228 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:00 crc kubenswrapper[4989]: E1006 09:03:00.929227 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667\": container with ID starting with 1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667 not found: ID does not exist" containerID="1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.929340 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667"} err="failed to get container status \"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667\": rpc error: code = NotFound desc = could not find container \"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667\": container with ID starting with 1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667 not found: ID does not exist" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.929369 4989 scope.go:117] "RemoveContainer" containerID="4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.929917 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7"} err="failed to get container status \"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7\": rpc error: code = NotFound desc = could not find container \"4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7\": container with ID starting with 4312bcee7297cbf3d2b61a91ba45f5cfe8d1dfac0b99f86064ddc0e816647eb7 not found: ID does not exist" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.929956 4989 scope.go:117] "RemoveContainer" containerID="1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.930447 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667"} err="failed to get container status \"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667\": rpc error: code = NotFound desc = could not find container \"1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667\": container with ID starting with 1224cfd07964d8b7000637fac4603cbc8dbfa2847f842b1c33324cc39a6d9667 not found: ID does not exist" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.932012 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.932293 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 09:03:00 crc kubenswrapper[4989]: I1006 09:03:00.937886 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.038546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.038866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlk87\" (UniqueName: \"kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.038961 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.039061 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.039159 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.141251 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlk87\" (UniqueName: \"kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.141748 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.141920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.142226 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.142462 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.143877 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.147005 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.147593 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.165404 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.170839 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlk87\" (UniqueName: \"kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87\") pod \"nova-metadata-0\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.266558 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.716191 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.858028 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerStarted","Data":"16743b7a768ea901fd6dac3a092f124d4d8bf9cfbb772ca09136297090292d5f"} Oct 06 09:03:01 crc kubenswrapper[4989]: I1006 09:03:01.954597 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d61f7286-65ea-4ed6-8d18-b4edf8afe373" path="/var/lib/kubelet/pods/d61f7286-65ea-4ed6-8d18-b4edf8afe373/volumes" Oct 06 09:03:02 crc kubenswrapper[4989]: I1006 09:03:02.870573 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerStarted","Data":"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c"} Oct 06 09:03:02 crc kubenswrapper[4989]: I1006 09:03:02.870940 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerStarted","Data":"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a"} Oct 06 09:03:02 crc kubenswrapper[4989]: I1006 09:03:02.899140 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.899115569 podStartE2EDuration="2.899115569s" podCreationTimestamp="2025-10-06 09:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:02.886642919 +0000 UTC m=+1433.676668499" watchObservedRunningTime="2025-10-06 09:03:02.899115569 +0000 UTC m=+1433.689141149" Oct 06 09:03:03 crc kubenswrapper[4989]: I1006 09:03:03.936067 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:03:03 crc kubenswrapper[4989]: I1006 09:03:03.936130 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:03:04 crc kubenswrapper[4989]: I1006 09:03:04.892923 4989 generic.go:334] "Generic (PLEG): container finished" podID="b416f76a-8dcd-411a-a0a0-31f7becea55e" containerID="58c5d959d29623aea9a09ac52e22bed81527dfc5edac480b7610c6595fb84cfd" exitCode=0 Oct 06 09:03:04 crc kubenswrapper[4989]: I1006 09:03:04.892967 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mwhkn" event={"ID":"b416f76a-8dcd-411a-a0a0-31f7becea55e","Type":"ContainerDied","Data":"58c5d959d29623aea9a09ac52e22bed81527dfc5edac480b7610c6595fb84cfd"} Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.202607 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.244793 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.322056 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.322127 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.444938 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.519393 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.519626 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="dnsmasq-dns" containerID="cri-o://27595d95b4570acc35bdf4a955f776adc14c0cb9a7b8a932b7cd599e5b47e33f" gracePeriod=10 Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.908526 4989 generic.go:334] "Generic (PLEG): container finished" podID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerID="27595d95b4570acc35bdf4a955f776adc14c0cb9a7b8a932b7cd599e5b47e33f" exitCode=0 Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.908744 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" event={"ID":"ccacf8e5-0fb0-41bd-92de-6851964f1ffb","Type":"ContainerDied","Data":"27595d95b4570acc35bdf4a955f776adc14c0cb9a7b8a932b7cd599e5b47e33f"} Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.912333 4989 generic.go:334] "Generic (PLEG): container finished" podID="8aec9ce1-0a3f-4788-a367-4fa2f1554d44" containerID="16a49375a3c8f1a4186ec2a1c51cfb2a9d9d97c9799db9d7fe6d33abe94affef" exitCode=0 Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.912419 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" event={"ID":"8aec9ce1-0a3f-4788-a367-4fa2f1554d44","Type":"ContainerDied","Data":"16a49375a3c8f1a4186ec2a1c51cfb2a9d9d97c9799db9d7fe6d33abe94affef"} Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.963555 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 09:03:05 crc kubenswrapper[4989]: I1006 09:03:05.969300 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.067726 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139458 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139523 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139640 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139824 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.139864 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl75r\" (UniqueName: \"kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r\") pod \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\" (UID: \"ccacf8e5-0fb0-41bd-92de-6851964f1ffb\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.188338 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r" (OuterVolumeSpecName: "kube-api-access-cl75r") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "kube-api-access-cl75r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.229212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.233623 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.242372 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.242411 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl75r\" (UniqueName: \"kubernetes.io/projected/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-kube-api-access-cl75r\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.242423 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.266630 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.267219 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.267426 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.267425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config" (OuterVolumeSpecName: "config") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.282708 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ccacf8e5-0fb0-41bd-92de-6851964f1ffb" (UID: "ccacf8e5-0fb0-41bd-92de-6851964f1ffb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.285957 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.344506 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.344550 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.344562 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ccacf8e5-0fb0-41bd-92de-6851964f1ffb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.404838 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.404839 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.446029 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcgg8\" (UniqueName: \"kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8\") pod \"b416f76a-8dcd-411a-a0a0-31f7becea55e\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.446149 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts\") pod \"b416f76a-8dcd-411a-a0a0-31f7becea55e\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.446241 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle\") pod \"b416f76a-8dcd-411a-a0a0-31f7becea55e\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.446306 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data\") pod \"b416f76a-8dcd-411a-a0a0-31f7becea55e\" (UID: \"b416f76a-8dcd-411a-a0a0-31f7becea55e\") " Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.450903 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts" (OuterVolumeSpecName: "scripts") pod "b416f76a-8dcd-411a-a0a0-31f7becea55e" (UID: "b416f76a-8dcd-411a-a0a0-31f7becea55e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.450924 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8" (OuterVolumeSpecName: "kube-api-access-tcgg8") pod "b416f76a-8dcd-411a-a0a0-31f7becea55e" (UID: "b416f76a-8dcd-411a-a0a0-31f7becea55e"). InnerVolumeSpecName "kube-api-access-tcgg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.480338 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data" (OuterVolumeSpecName: "config-data") pod "b416f76a-8dcd-411a-a0a0-31f7becea55e" (UID: "b416f76a-8dcd-411a-a0a0-31f7becea55e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.482961 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b416f76a-8dcd-411a-a0a0-31f7becea55e" (UID: "b416f76a-8dcd-411a-a0a0-31f7becea55e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.548921 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.548953 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.548962 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcgg8\" (UniqueName: \"kubernetes.io/projected/b416f76a-8dcd-411a-a0a0-31f7becea55e-kube-api-access-tcgg8\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.548971 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b416f76a-8dcd-411a-a0a0-31f7becea55e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.921955 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" event={"ID":"ccacf8e5-0fb0-41bd-92de-6851964f1ffb","Type":"ContainerDied","Data":"846a3187970586e24834545824db2dfa823583f4447875d1527e56f4827bfb51"} Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.922013 4989 scope.go:117] "RemoveContainer" containerID="27595d95b4570acc35bdf4a955f776adc14c0cb9a7b8a932b7cd599e5b47e33f" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.922150 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-tjzqg" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.928935 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mwhkn" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.931766 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mwhkn" event={"ID":"b416f76a-8dcd-411a-a0a0-31f7becea55e","Type":"ContainerDied","Data":"e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062"} Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.931834 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4b550892f79dff36ba6b2cf6bc436792827a369c360685c800ee661436f3062" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.981060 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.994573 4989 scope.go:117] "RemoveContainer" containerID="bc52c255a8e5d24ea77357689e284842bed6b4d5af90089348952e7a7ce407cf" Oct 06 09:03:06 crc kubenswrapper[4989]: I1006 09:03:06.997000 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-tjzqg"] Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.101461 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.101682 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-log" containerID="cri-o://26863d5b0d9b2d53be9b4e56d60c8e66f73ac31b5497f62f6b98628cde8cac5d" gracePeriod=30 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.101810 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-api" containerID="cri-o://ebb72deedd186195f6b578c0c2fa80fae603ae9607847a578e8bb536f1883b67" gracePeriod=30 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.115519 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.125780 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.324516 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.473262 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data\") pod \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.473678 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2ftv\" (UniqueName: \"kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv\") pod \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.473901 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle\") pod \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.474053 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts\") pod \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\" (UID: \"8aec9ce1-0a3f-4788-a367-4fa2f1554d44\") " Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.499896 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv" (OuterVolumeSpecName: "kube-api-access-d2ftv") pod "8aec9ce1-0a3f-4788-a367-4fa2f1554d44" (UID: "8aec9ce1-0a3f-4788-a367-4fa2f1554d44"). InnerVolumeSpecName "kube-api-access-d2ftv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.502074 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts" (OuterVolumeSpecName: "scripts") pod "8aec9ce1-0a3f-4788-a367-4fa2f1554d44" (UID: "8aec9ce1-0a3f-4788-a367-4fa2f1554d44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.507060 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aec9ce1-0a3f-4788-a367-4fa2f1554d44" (UID: "8aec9ce1-0a3f-4788-a367-4fa2f1554d44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.507583 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data" (OuterVolumeSpecName: "config-data") pod "8aec9ce1-0a3f-4788-a367-4fa2f1554d44" (UID: "8aec9ce1-0a3f-4788-a367-4fa2f1554d44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.576972 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.577556 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.577746 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.577879 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2ftv\" (UniqueName: \"kubernetes.io/projected/8aec9ce1-0a3f-4788-a367-4fa2f1554d44-kube-api-access-d2ftv\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.944257 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.950092 4989 generic.go:334] "Generic (PLEG): container finished" podID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerID="26863d5b0d9b2d53be9b4e56d60c8e66f73ac31b5497f62f6b98628cde8cac5d" exitCode=143 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.950340 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f4386782-615c-4aa0-960f-83c312cbb724" containerName="nova-scheduler-scheduler" containerID="cri-o://08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" gracePeriod=30 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.950604 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-log" containerID="cri-o://6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" gracePeriod=30 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.950737 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-metadata" containerID="cri-o://1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" gracePeriod=30 Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.974221 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" path="/var/lib/kubelet/pods/ccacf8e5-0fb0-41bd-92de-6851964f1ffb/volumes" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.980154 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-r7w8z" event={"ID":"8aec9ce1-0a3f-4788-a367-4fa2f1554d44","Type":"ContainerDied","Data":"33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243"} Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.980216 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33eda37b37da832756c354c04139c553c4f23478535c5fbb05c292ec8a270243" Oct 06 09:03:07 crc kubenswrapper[4989]: I1006 09:03:07.980245 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerDied","Data":"26863d5b0d9b2d53be9b4e56d60c8e66f73ac31b5497f62f6b98628cde8cac5d"} Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041151 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:03:08 crc kubenswrapper[4989]: E1006 09:03:08.041526 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="dnsmasq-dns" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041539 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="dnsmasq-dns" Oct 06 09:03:08 crc kubenswrapper[4989]: E1006 09:03:08.041554 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="init" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041559 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="init" Oct 06 09:03:08 crc kubenswrapper[4989]: E1006 09:03:08.041585 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aec9ce1-0a3f-4788-a367-4fa2f1554d44" containerName="nova-cell1-conductor-db-sync" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041592 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aec9ce1-0a3f-4788-a367-4fa2f1554d44" containerName="nova-cell1-conductor-db-sync" Oct 06 09:03:08 crc kubenswrapper[4989]: E1006 09:03:08.041601 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b416f76a-8dcd-411a-a0a0-31f7becea55e" containerName="nova-manage" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041607 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b416f76a-8dcd-411a-a0a0-31f7becea55e" containerName="nova-manage" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041802 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aec9ce1-0a3f-4788-a367-4fa2f1554d44" containerName="nova-cell1-conductor-db-sync" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041816 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b416f76a-8dcd-411a-a0a0-31f7becea55e" containerName="nova-manage" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.041839 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccacf8e5-0fb0-41bd-92de-6851964f1ffb" containerName="dnsmasq-dns" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.042444 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.052313 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.052872 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.200722 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.200776 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plr2k\" (UniqueName: \"kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.200819 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.302469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.302524 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plr2k\" (UniqueName: \"kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.302557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.307452 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.311574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.326526 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plr2k\" (UniqueName: \"kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k\") pod \"nova-cell1-conductor-0\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.436375 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.860826 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.920735 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.960403 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73314917-dc1f-470c-b2d6-932ebcd443b8","Type":"ContainerStarted","Data":"f3f823bab6661fa109dab69c29d53e8bb7313b585270c503620ce968978c078f"} Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963538 4989 generic.go:334] "Generic (PLEG): container finished" podID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerID="1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" exitCode=0 Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963565 4989 generic.go:334] "Generic (PLEG): container finished" podID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerID="6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" exitCode=143 Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963584 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerDied","Data":"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c"} Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963604 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963623 4989 scope.go:117] "RemoveContainer" containerID="1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerDied","Data":"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a"} Oct 06 09:03:08 crc kubenswrapper[4989]: I1006 09:03:08.963776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"33a2d59c-4a0c-48a0-a076-29d3a3953b87","Type":"ContainerDied","Data":"16743b7a768ea901fd6dac3a092f124d4d8bf9cfbb772ca09136297090292d5f"} Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.014247 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs\") pod \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.014308 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle\") pod \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.014424 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data\") pod \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.014527 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs\") pod \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.014553 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlk87\" (UniqueName: \"kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87\") pod \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\" (UID: \"33a2d59c-4a0c-48a0-a076-29d3a3953b87\") " Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.015082 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs" (OuterVolumeSpecName: "logs") pod "33a2d59c-4a0c-48a0-a076-29d3a3953b87" (UID: "33a2d59c-4a0c-48a0-a076-29d3a3953b87"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.019845 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87" (OuterVolumeSpecName: "kube-api-access-mlk87") pod "33a2d59c-4a0c-48a0-a076-29d3a3953b87" (UID: "33a2d59c-4a0c-48a0-a076-29d3a3953b87"). InnerVolumeSpecName "kube-api-access-mlk87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.042712 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33a2d59c-4a0c-48a0-a076-29d3a3953b87" (UID: "33a2d59c-4a0c-48a0-a076-29d3a3953b87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.043971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data" (OuterVolumeSpecName: "config-data") pod "33a2d59c-4a0c-48a0-a076-29d3a3953b87" (UID: "33a2d59c-4a0c-48a0-a076-29d3a3953b87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.057933 4989 scope.go:117] "RemoveContainer" containerID="6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.081993 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "33a2d59c-4a0c-48a0-a076-29d3a3953b87" (UID: "33a2d59c-4a0c-48a0-a076-29d3a3953b87"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.082106 4989 scope.go:117] "RemoveContainer" containerID="1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" Oct 06 09:03:09 crc kubenswrapper[4989]: E1006 09:03:09.082525 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c\": container with ID starting with 1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c not found: ID does not exist" containerID="1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.082573 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c"} err="failed to get container status \"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c\": rpc error: code = NotFound desc = could not find container \"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c\": container with ID starting with 1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c not found: ID does not exist" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.082602 4989 scope.go:117] "RemoveContainer" containerID="6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" Oct 06 09:03:09 crc kubenswrapper[4989]: E1006 09:03:09.083068 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a\": container with ID starting with 6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a not found: ID does not exist" containerID="6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.083094 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a"} err="failed to get container status \"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a\": rpc error: code = NotFound desc = could not find container \"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a\": container with ID starting with 6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a not found: ID does not exist" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.083109 4989 scope.go:117] "RemoveContainer" containerID="1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.083465 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c"} err="failed to get container status \"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c\": rpc error: code = NotFound desc = could not find container \"1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c\": container with ID starting with 1e3219804de7aae41441c23a804e955a3e3a7c821d8d00cac17635330c98210c not found: ID does not exist" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.083514 4989 scope.go:117] "RemoveContainer" containerID="6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.083842 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a"} err="failed to get container status \"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a\": rpc error: code = NotFound desc = could not find container \"6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a\": container with ID starting with 6c8382c41c8dc7ab54c20ff99e4fc62476c78e560b9623e0f06a64c75df29d1a not found: ID does not exist" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.117156 4989 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.117186 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.117197 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33a2d59c-4a0c-48a0-a076-29d3a3953b87-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.117205 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/33a2d59c-4a0c-48a0-a076-29d3a3953b87-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.117214 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlk87\" (UniqueName: \"kubernetes.io/projected/33a2d59c-4a0c-48a0-a076-29d3a3953b87-kube-api-access-mlk87\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.298181 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.307906 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.318948 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:09 crc kubenswrapper[4989]: E1006 09:03:09.319525 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-metadata" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.319594 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-metadata" Oct 06 09:03:09 crc kubenswrapper[4989]: E1006 09:03:09.319729 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-log" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.319786 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-log" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.320023 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-log" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.320096 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" containerName="nova-metadata-metadata" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.321504 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.325301 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.325406 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.336523 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.421649 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.421743 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.421765 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.421867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvs2d\" (UniqueName: \"kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.421923 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.523517 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.524020 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.525167 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.525483 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvs2d\" (UniqueName: \"kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.525825 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.526128 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.528133 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.528369 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.529465 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.544819 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvs2d\" (UniqueName: \"kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d\") pod \"nova-metadata-0\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.638224 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.962283 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33a2d59c-4a0c-48a0-a076-29d3a3953b87" path="/var/lib/kubelet/pods/33a2d59c-4a0c-48a0-a076-29d3a3953b87/volumes" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.974977 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73314917-dc1f-470c-b2d6-932ebcd443b8","Type":"ContainerStarted","Data":"a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe"} Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.975730 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:09 crc kubenswrapper[4989]: I1006 09:03:09.999075 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.999053464 podStartE2EDuration="1.999053464s" podCreationTimestamp="2025-10-06 09:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:09.995942614 +0000 UTC m=+1440.785968194" watchObservedRunningTime="2025-10-06 09:03:09.999053464 +0000 UTC m=+1440.789079044" Oct 06 09:03:10 crc kubenswrapper[4989]: I1006 09:03:10.171177 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:10 crc kubenswrapper[4989]: E1006 09:03:10.206408 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:10 crc kubenswrapper[4989]: E1006 09:03:10.214307 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:10 crc kubenswrapper[4989]: E1006 09:03:10.233821 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:10 crc kubenswrapper[4989]: E1006 09:03:10.233910 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f4386782-615c-4aa0-960f-83c312cbb724" containerName="nova-scheduler-scheduler" Oct 06 09:03:10 crc kubenswrapper[4989]: I1006 09:03:10.994447 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerStarted","Data":"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3"} Oct 06 09:03:10 crc kubenswrapper[4989]: I1006 09:03:10.994936 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerStarted","Data":"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5"} Oct 06 09:03:10 crc kubenswrapper[4989]: I1006 09:03:10.994950 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerStarted","Data":"0897d1eeeb60ab92ac6614f1dc3ab345644a6c768024308f83211bd634387ed0"} Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.046552 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.046527713 podStartE2EDuration="2.046527713s" podCreationTimestamp="2025-10-06 09:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:11.039869111 +0000 UTC m=+1441.829894701" watchObservedRunningTime="2025-10-06 09:03:11.046527713 +0000 UTC m=+1441.836553293" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.782587 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.885551 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle\") pod \"f4386782-615c-4aa0-960f-83c312cbb724\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.885926 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data\") pod \"f4386782-615c-4aa0-960f-83c312cbb724\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.886043 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dwqq\" (UniqueName: \"kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq\") pod \"f4386782-615c-4aa0-960f-83c312cbb724\" (UID: \"f4386782-615c-4aa0-960f-83c312cbb724\") " Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.904980 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq" (OuterVolumeSpecName: "kube-api-access-7dwqq") pod "f4386782-615c-4aa0-960f-83c312cbb724" (UID: "f4386782-615c-4aa0-960f-83c312cbb724"). InnerVolumeSpecName "kube-api-access-7dwqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.912565 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data" (OuterVolumeSpecName: "config-data") pod "f4386782-615c-4aa0-960f-83c312cbb724" (UID: "f4386782-615c-4aa0-960f-83c312cbb724"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.922124 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4386782-615c-4aa0-960f-83c312cbb724" (UID: "f4386782-615c-4aa0-960f-83c312cbb724"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.988099 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dwqq\" (UniqueName: \"kubernetes.io/projected/f4386782-615c-4aa0-960f-83c312cbb724-kube-api-access-7dwqq\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.988134 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:11 crc kubenswrapper[4989]: I1006 09:03:11.988143 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4386782-615c-4aa0-960f-83c312cbb724-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.006313 4989 generic.go:334] "Generic (PLEG): container finished" podID="f4386782-615c-4aa0-960f-83c312cbb724" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" exitCode=0 Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.006376 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4386782-615c-4aa0-960f-83c312cbb724","Type":"ContainerDied","Data":"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc"} Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.006414 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.006445 4989 scope.go:117] "RemoveContainer" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.006432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f4386782-615c-4aa0-960f-83c312cbb724","Type":"ContainerDied","Data":"baef450d5c104aa1305dcffacc6be9bdbf6a983c93211f900f7dd8f78ae03c1d"} Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.029971 4989 scope.go:117] "RemoveContainer" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.030557 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:12 crc kubenswrapper[4989]: E1006 09:03:12.030828 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc\": container with ID starting with 08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc not found: ID does not exist" containerID="08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.030881 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc"} err="failed to get container status \"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc\": rpc error: code = NotFound desc = could not find container \"08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc\": container with ID starting with 08a99290806395853d427f888951b1ff82342a87f893ff94bdcf351f9d3728cc not found: ID does not exist" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.040016 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.051041 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:12 crc kubenswrapper[4989]: E1006 09:03:12.051436 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4386782-615c-4aa0-960f-83c312cbb724" containerName="nova-scheduler-scheduler" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.051448 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4386782-615c-4aa0-960f-83c312cbb724" containerName="nova-scheduler-scheduler" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.051603 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4386782-615c-4aa0-960f-83c312cbb724" containerName="nova-scheduler-scheduler" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.052346 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.056008 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.062243 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.190872 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.191247 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.191464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc9kh\" (UniqueName: \"kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.294148 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.294344 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.294428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc9kh\" (UniqueName: \"kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.299106 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.299167 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.314254 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc9kh\" (UniqueName: \"kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh\") pod \"nova-scheduler-0\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.402687 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:12 crc kubenswrapper[4989]: I1006 09:03:12.857423 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.017077 4989 generic.go:334] "Generic (PLEG): container finished" podID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerID="ebb72deedd186195f6b578c0c2fa80fae603ae9607847a578e8bb536f1883b67" exitCode=0 Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.017116 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerDied","Data":"ebb72deedd186195f6b578c0c2fa80fae603ae9607847a578e8bb536f1883b67"} Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.017394 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ffc66475-afda-455f-b283-cbcc213fb7fc","Type":"ContainerDied","Data":"ab7a9f16c03f909d5a07a3b955c22bb5d3e28beff36c00d32d86b61f1e9ee8fd"} Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.017408 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab7a9f16c03f909d5a07a3b955c22bb5d3e28beff36c00d32d86b61f1e9ee8fd" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.020754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aeb4f1b9-106e-42a4-93b1-292448b7e557","Type":"ContainerStarted","Data":"f649817a9f7ee0ca05c7720d837459a76ea2d0868546d3044e5d87c4420f4bb4"} Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.030525 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.114294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs\") pod \"ffc66475-afda-455f-b283-cbcc213fb7fc\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.114619 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") pod \"ffc66475-afda-455f-b283-cbcc213fb7fc\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.114797 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle\") pod \"ffc66475-afda-455f-b283-cbcc213fb7fc\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.114920 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs" (OuterVolumeSpecName: "logs") pod "ffc66475-afda-455f-b283-cbcc213fb7fc" (UID: "ffc66475-afda-455f-b283-cbcc213fb7fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.115145 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngcf\" (UniqueName: \"kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf\") pod \"ffc66475-afda-455f-b283-cbcc213fb7fc\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.115907 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ffc66475-afda-455f-b283-cbcc213fb7fc-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.118411 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf" (OuterVolumeSpecName: "kube-api-access-cngcf") pod "ffc66475-afda-455f-b283-cbcc213fb7fc" (UID: "ffc66475-afda-455f-b283-cbcc213fb7fc"). InnerVolumeSpecName "kube-api-access-cngcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:13 crc kubenswrapper[4989]: E1006 09:03:13.142484 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data podName:ffc66475-afda-455f-b283-cbcc213fb7fc nodeName:}" failed. No retries permitted until 2025-10-06 09:03:13.642450021 +0000 UTC m=+1444.432475601 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data") pod "ffc66475-afda-455f-b283-cbcc213fb7fc" (UID: "ffc66475-afda-455f-b283-cbcc213fb7fc") : error deleting /var/lib/kubelet/pods/ffc66475-afda-455f-b283-cbcc213fb7fc/volume-subpaths: remove /var/lib/kubelet/pods/ffc66475-afda-455f-b283-cbcc213fb7fc/volume-subpaths: no such file or directory Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.145620 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffc66475-afda-455f-b283-cbcc213fb7fc" (UID: "ffc66475-afda-455f-b283-cbcc213fb7fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.218014 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngcf\" (UniqueName: \"kubernetes.io/projected/ffc66475-afda-455f-b283-cbcc213fb7fc-kube-api-access-cngcf\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.218043 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.727328 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") pod \"ffc66475-afda-455f-b283-cbcc213fb7fc\" (UID: \"ffc66475-afda-455f-b283-cbcc213fb7fc\") " Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.738304 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data" (OuterVolumeSpecName: "config-data") pod "ffc66475-afda-455f-b283-cbcc213fb7fc" (UID: "ffc66475-afda-455f-b283-cbcc213fb7fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.829588 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffc66475-afda-455f-b283-cbcc213fb7fc-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:13 crc kubenswrapper[4989]: I1006 09:03:13.946155 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4386782-615c-4aa0-960f-83c312cbb724" path="/var/lib/kubelet/pods/f4386782-615c-4aa0-960f-83c312cbb724/volumes" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.029768 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.031082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aeb4f1b9-106e-42a4-93b1-292448b7e557","Type":"ContainerStarted","Data":"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0"} Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.051594 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.05157806 podStartE2EDuration="2.05157806s" podCreationTimestamp="2025-10-06 09:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:14.048495031 +0000 UTC m=+1444.838520601" watchObservedRunningTime="2025-10-06 09:03:14.05157806 +0000 UTC m=+1444.841603640" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.076028 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.096244 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.110569 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:14 crc kubenswrapper[4989]: E1006 09:03:14.110998 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-log" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.111014 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-log" Oct 06 09:03:14 crc kubenswrapper[4989]: E1006 09:03:14.111048 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-api" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.111055 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-api" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.111275 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-log" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.111298 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" containerName="nova-api-api" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.112440 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.114698 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.123812 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.238888 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.239081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.239188 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.239268 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r98zv\" (UniqueName: \"kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.341527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.341581 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.341619 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.341694 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r98zv\" (UniqueName: \"kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.342589 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.355064 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.357030 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.367187 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r98zv\" (UniqueName: \"kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv\") pod \"nova-api-0\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.440173 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.639441 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.639836 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:03:14 crc kubenswrapper[4989]: I1006 09:03:14.911980 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:14 crc kubenswrapper[4989]: W1006 09:03:14.914431 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod461d00bd_1c96_4406_a64a_32f159af8a1b.slice/crio-0ef058e4e993d253d6f0f0af691f7a349d9891b16af050a3183a6fa4b0a278d7 WatchSource:0}: Error finding container 0ef058e4e993d253d6f0f0af691f7a349d9891b16af050a3183a6fa4b0a278d7: Status 404 returned error can't find the container with id 0ef058e4e993d253d6f0f0af691f7a349d9891b16af050a3183a6fa4b0a278d7 Oct 06 09:03:15 crc kubenswrapper[4989]: I1006 09:03:15.041149 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerStarted","Data":"0ef058e4e993d253d6f0f0af691f7a349d9891b16af050a3183a6fa4b0a278d7"} Oct 06 09:03:15 crc kubenswrapper[4989]: I1006 09:03:15.948911 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc66475-afda-455f-b283-cbcc213fb7fc" path="/var/lib/kubelet/pods/ffc66475-afda-455f-b283-cbcc213fb7fc/volumes" Oct 06 09:03:16 crc kubenswrapper[4989]: I1006 09:03:16.052556 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerStarted","Data":"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d"} Oct 06 09:03:16 crc kubenswrapper[4989]: I1006 09:03:16.052608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerStarted","Data":"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4"} Oct 06 09:03:16 crc kubenswrapper[4989]: I1006 09:03:16.083906 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.083883043 podStartE2EDuration="2.083883043s" podCreationTimestamp="2025-10-06 09:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:16.074001078 +0000 UTC m=+1446.864026678" watchObservedRunningTime="2025-10-06 09:03:16.083883043 +0000 UTC m=+1446.873908643" Oct 06 09:03:17 crc kubenswrapper[4989]: I1006 09:03:17.403371 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 09:03:18 crc kubenswrapper[4989]: I1006 09:03:18.475561 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 06 09:03:19 crc kubenswrapper[4989]: I1006 09:03:19.639141 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 09:03:19 crc kubenswrapper[4989]: I1006 09:03:19.639266 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 09:03:20 crc kubenswrapper[4989]: I1006 09:03:20.652972 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:20 crc kubenswrapper[4989]: I1006 09:03:20.653531 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:22 crc kubenswrapper[4989]: I1006 09:03:22.403227 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 09:03:22 crc kubenswrapper[4989]: I1006 09:03:22.435174 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 09:03:23 crc kubenswrapper[4989]: I1006 09:03:23.134387 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 09:03:24 crc kubenswrapper[4989]: I1006 09:03:24.440943 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:03:24 crc kubenswrapper[4989]: I1006 09:03:24.441011 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.496871 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.499889 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.517551 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.524920 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.524929 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.660896 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.661058 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jdmf\" (UniqueName: \"kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.661171 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.762935 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.763011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jdmf\" (UniqueName: \"kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.763047 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.763495 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.763523 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.798785 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jdmf\" (UniqueName: \"kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf\") pod \"redhat-marketplace-wdwgx\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:25 crc kubenswrapper[4989]: I1006 09:03:25.820422 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:26 crc kubenswrapper[4989]: I1006 09:03:26.295348 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:26 crc kubenswrapper[4989]: W1006 09:03:26.302379 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ddc1aa3_4dd2_4208_a199_2b867a42968f.slice/crio-35483b361b01e4297f04197f2b7a957c8e595d3f65664524c68c4fdd23a5af5d WatchSource:0}: Error finding container 35483b361b01e4297f04197f2b7a957c8e595d3f65664524c68c4fdd23a5af5d: Status 404 returned error can't find the container with id 35483b361b01e4297f04197f2b7a957c8e595d3f65664524c68c4fdd23a5af5d Oct 06 09:03:27 crc kubenswrapper[4989]: I1006 09:03:27.142331 4989 generic.go:334] "Generic (PLEG): container finished" podID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerID="d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637" exitCode=0 Oct 06 09:03:27 crc kubenswrapper[4989]: I1006 09:03:27.142423 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerDied","Data":"d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637"} Oct 06 09:03:27 crc kubenswrapper[4989]: I1006 09:03:27.142626 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerStarted","Data":"35483b361b01e4297f04197f2b7a957c8e595d3f65664524c68c4fdd23a5af5d"} Oct 06 09:03:29 crc kubenswrapper[4989]: I1006 09:03:29.161804 4989 generic.go:334] "Generic (PLEG): container finished" podID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerID="85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0" exitCode=0 Oct 06 09:03:29 crc kubenswrapper[4989]: I1006 09:03:29.161860 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerDied","Data":"85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0"} Oct 06 09:03:29 crc kubenswrapper[4989]: I1006 09:03:29.646492 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 09:03:29 crc kubenswrapper[4989]: I1006 09:03:29.651350 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 09:03:29 crc kubenswrapper[4989]: I1006 09:03:29.653174 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.173481 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerStarted","Data":"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62"} Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.176037 4989 generic.go:334] "Generic (PLEG): container finished" podID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" containerID="43c2e169810eb04bada5f35eaaf81350cd70184b618b76a271f06ac3f1d73440" exitCode=137 Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.177249 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"377dacd9-0ff0-4fab-8cbd-e28f205efb0d","Type":"ContainerDied","Data":"43c2e169810eb04bada5f35eaaf81350cd70184b618b76a271f06ac3f1d73440"} Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.177369 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"377dacd9-0ff0-4fab-8cbd-e28f205efb0d","Type":"ContainerDied","Data":"781f6c70c9481c574bb4fa7d81df54dd605c61718e54f5d1ae2c126620b0d0ac"} Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.177424 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="781f6c70c9481c574bb4fa7d81df54dd605c61718e54f5d1ae2c126620b0d0ac" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.188064 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.199237 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wdwgx" podStartSLOduration=2.650951322 podStartE2EDuration="5.199214054s" podCreationTimestamp="2025-10-06 09:03:25 +0000 UTC" firstStartedPulling="2025-10-06 09:03:27.145242096 +0000 UTC m=+1457.935267666" lastFinishedPulling="2025-10-06 09:03:29.693504818 +0000 UTC m=+1460.483530398" observedRunningTime="2025-10-06 09:03:30.190244695 +0000 UTC m=+1460.980270275" watchObservedRunningTime="2025-10-06 09:03:30.199214054 +0000 UTC m=+1460.989239634" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.269465 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.445226 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data\") pod \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.445399 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle\") pod \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.445462 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkmll\" (UniqueName: \"kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll\") pod \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\" (UID: \"377dacd9-0ff0-4fab-8cbd-e28f205efb0d\") " Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.450582 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll" (OuterVolumeSpecName: "kube-api-access-zkmll") pod "377dacd9-0ff0-4fab-8cbd-e28f205efb0d" (UID: "377dacd9-0ff0-4fab-8cbd-e28f205efb0d"). InnerVolumeSpecName "kube-api-access-zkmll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.475848 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data" (OuterVolumeSpecName: "config-data") pod "377dacd9-0ff0-4fab-8cbd-e28f205efb0d" (UID: "377dacd9-0ff0-4fab-8cbd-e28f205efb0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.499784 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "377dacd9-0ff0-4fab-8cbd-e28f205efb0d" (UID: "377dacd9-0ff0-4fab-8cbd-e28f205efb0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.547275 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkmll\" (UniqueName: \"kubernetes.io/projected/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-kube-api-access-zkmll\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.547322 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:30 crc kubenswrapper[4989]: I1006 09:03:30.547336 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377dacd9-0ff0-4fab-8cbd-e28f205efb0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.191490 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.232639 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.245284 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.264532 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:03:31 crc kubenswrapper[4989]: E1006 09:03:31.265142 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.265170 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.265532 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.266859 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.274136 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.274375 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.274597 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.278131 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.365478 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.365756 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdn9\" (UniqueName: \"kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.366222 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.366389 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.366484 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.468598 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdn9\" (UniqueName: \"kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.468748 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.468781 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.468812 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.468882 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.473458 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.473965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.474170 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.474996 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.488804 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdn9\" (UniqueName: \"kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.599617 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:31 crc kubenswrapper[4989]: I1006 09:03:31.948006 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="377dacd9-0ff0-4fab-8cbd-e28f205efb0d" path="/var/lib/kubelet/pods/377dacd9-0ff0-4fab-8cbd-e28f205efb0d/volumes" Oct 06 09:03:32 crc kubenswrapper[4989]: I1006 09:03:32.032450 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:03:32 crc kubenswrapper[4989]: I1006 09:03:32.200135 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50fc7e8a-a4c4-43ca-b672-7d8f63e25738","Type":"ContainerStarted","Data":"056cd8afe444128c40113c0f59d0428a85aa64ee07b6b59a9d92282bfabe65cb"} Oct 06 09:03:33 crc kubenswrapper[4989]: I1006 09:03:33.212134 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50fc7e8a-a4c4-43ca-b672-7d8f63e25738","Type":"ContainerStarted","Data":"fa81e6c26183ea833e5e2d25e80e937b2130d7c82f3b53456b03c4b55db2e255"} Oct 06 09:03:33 crc kubenswrapper[4989]: I1006 09:03:33.234713 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.234691787 podStartE2EDuration="2.234691787s" podCreationTimestamp="2025-10-06 09:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:33.231842035 +0000 UTC m=+1464.021867615" watchObservedRunningTime="2025-10-06 09:03:33.234691787 +0000 UTC m=+1464.024717367" Oct 06 09:03:33 crc kubenswrapper[4989]: I1006 09:03:33.938755 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:03:33 crc kubenswrapper[4989]: I1006 09:03:33.938870 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:03:34 crc kubenswrapper[4989]: I1006 09:03:34.444959 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 09:03:34 crc kubenswrapper[4989]: I1006 09:03:34.446308 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 09:03:34 crc kubenswrapper[4989]: I1006 09:03:34.446510 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 09:03:34 crc kubenswrapper[4989]: I1006 09:03:34.450635 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.228787 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.231689 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.409959 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.412847 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.436364 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.589894 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.589957 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66gtt\" (UniqueName: \"kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.589977 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.590018 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.590058 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.590106 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.691814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.692768 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.692942 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.693600 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.693874 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.694462 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.694544 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66gtt\" (UniqueName: \"kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.694904 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.695583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.695762 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.696483 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.715416 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66gtt\" (UniqueName: \"kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt\") pod \"dnsmasq-dns-59cf4bdb65-bmrx4\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.741567 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.820986 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.822117 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:35 crc kubenswrapper[4989]: I1006 09:03:35.905382 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:36 crc kubenswrapper[4989]: I1006 09:03:36.291703 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:03:36 crc kubenswrapper[4989]: I1006 09:03:36.310120 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:36 crc kubenswrapper[4989]: I1006 09:03:36.370855 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:36 crc kubenswrapper[4989]: I1006 09:03:36.600564 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.249673 4989 generic.go:334] "Generic (PLEG): container finished" podID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerID="9be8370edf964b02bc2eec83c55e0ec192884f6e6390d307a5c436a87e7f3ff2" exitCode=0 Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.250041 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" event={"ID":"80aa68cc-125e-4753-8c49-17f39da3e3f5","Type":"ContainerDied","Data":"9be8370edf964b02bc2eec83c55e0ec192884f6e6390d307a5c436a87e7f3ff2"} Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.250069 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" event={"ID":"80aa68cc-125e-4753-8c49-17f39da3e3f5","Type":"ContainerStarted","Data":"a67edcbcec88bf76f64b2e568efe9e42177d529c33d9a9979cd05231aed7ef2d"} Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.695423 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.696128 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-central-agent" containerID="cri-o://182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d" gracePeriod=30 Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.696672 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="proxy-httpd" containerID="cri-o://0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07" gracePeriod=30 Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.696820 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="sg-core" containerID="cri-o://e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208" gracePeriod=30 Oct 06 09:03:37 crc kubenswrapper[4989]: I1006 09:03:37.696837 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-notification-agent" containerID="cri-o://901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30" gracePeriod=30 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.127691 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.260976 4989 generic.go:334] "Generic (PLEG): container finished" podID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerID="0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07" exitCode=0 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.261010 4989 generic.go:334] "Generic (PLEG): container finished" podID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerID="e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208" exitCode=2 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.261020 4989 generic.go:334] "Generic (PLEG): container finished" podID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerID="182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d" exitCode=0 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.261033 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerDied","Data":"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07"} Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.261071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerDied","Data":"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208"} Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.261083 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerDied","Data":"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d"} Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.263132 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" event={"ID":"80aa68cc-125e-4753-8c49-17f39da3e3f5","Type":"ContainerStarted","Data":"57bccaac485a945bfa34ccfa394125f43a85083ec91573de318dfac2c1dbab2e"} Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.263306 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wdwgx" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="registry-server" containerID="cri-o://44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62" gracePeriod=2 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.263400 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-log" containerID="cri-o://8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4" gracePeriod=30 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.263458 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-api" containerID="cri-o://88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d" gracePeriod=30 Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.287076 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" podStartSLOduration=3.287053329 podStartE2EDuration="3.287053329s" podCreationTimestamp="2025-10-06 09:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:38.281292233 +0000 UTC m=+1469.071317813" watchObservedRunningTime="2025-10-06 09:03:38.287053329 +0000 UTC m=+1469.077078909" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.667441 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.757762 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content\") pod \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.757891 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jdmf\" (UniqueName: \"kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf\") pod \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.758029 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities\") pod \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\" (UID: \"2ddc1aa3-4dd2-4208-a199-2b867a42968f\") " Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.760360 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities" (OuterVolumeSpecName: "utilities") pod "2ddc1aa3-4dd2-4208-a199-2b867a42968f" (UID: "2ddc1aa3-4dd2-4208-a199-2b867a42968f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.764395 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf" (OuterVolumeSpecName: "kube-api-access-7jdmf") pod "2ddc1aa3-4dd2-4208-a199-2b867a42968f" (UID: "2ddc1aa3-4dd2-4208-a199-2b867a42968f"). InnerVolumeSpecName "kube-api-access-7jdmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.770929 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ddc1aa3-4dd2-4208-a199-2b867a42968f" (UID: "2ddc1aa3-4dd2-4208-a199-2b867a42968f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.860294 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.860329 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ddc1aa3-4dd2-4208-a199-2b867a42968f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:38 crc kubenswrapper[4989]: I1006 09:03:38.860340 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jdmf\" (UniqueName: \"kubernetes.io/projected/2ddc1aa3-4dd2-4208-a199-2b867a42968f-kube-api-access-7jdmf\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.273696 4989 generic.go:334] "Generic (PLEG): container finished" podID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerID="8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4" exitCode=143 Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.273759 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerDied","Data":"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4"} Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276005 4989 generic.go:334] "Generic (PLEG): container finished" podID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerID="44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62" exitCode=0 Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276056 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wdwgx" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276076 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerDied","Data":"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62"} Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276115 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wdwgx" event={"ID":"2ddc1aa3-4dd2-4208-a199-2b867a42968f","Type":"ContainerDied","Data":"35483b361b01e4297f04197f2b7a957c8e595d3f65664524c68c4fdd23a5af5d"} Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276134 4989 scope.go:117] "RemoveContainer" containerID="44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.276997 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.304889 4989 scope.go:117] "RemoveContainer" containerID="85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.307638 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.333853 4989 scope.go:117] "RemoveContainer" containerID="d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.336854 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wdwgx"] Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.374122 4989 scope.go:117] "RemoveContainer" containerID="44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62" Oct 06 09:03:39 crc kubenswrapper[4989]: E1006 09:03:39.374466 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62\": container with ID starting with 44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62 not found: ID does not exist" containerID="44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.374496 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62"} err="failed to get container status \"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62\": rpc error: code = NotFound desc = could not find container \"44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62\": container with ID starting with 44cedf38105c2d659dbea1f13d7c00666a396c33cc05802e08a1cb112c257d62 not found: ID does not exist" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.374516 4989 scope.go:117] "RemoveContainer" containerID="85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0" Oct 06 09:03:39 crc kubenswrapper[4989]: E1006 09:03:39.374739 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0\": container with ID starting with 85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0 not found: ID does not exist" containerID="85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.374765 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0"} err="failed to get container status \"85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0\": rpc error: code = NotFound desc = could not find container \"85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0\": container with ID starting with 85cfbf3ddc68bcf8934e04c1ad6251851aee23945542aeb6c5705228852c7ae0 not found: ID does not exist" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.374783 4989 scope.go:117] "RemoveContainer" containerID="d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637" Oct 06 09:03:39 crc kubenswrapper[4989]: E1006 09:03:39.374996 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637\": container with ID starting with d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637 not found: ID does not exist" containerID="d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.375020 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637"} err="failed to get container status \"d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637\": rpc error: code = NotFound desc = could not find container \"d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637\": container with ID starting with d69e106429909a1322354e214623621b2e15b41bce352c5a1e9be845cbde8637 not found: ID does not exist" Oct 06 09:03:39 crc kubenswrapper[4989]: I1006 09:03:39.945971 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" path="/var/lib/kubelet/pods/2ddc1aa3-4dd2-4208-a199-2b867a42968f/volumes" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.742036 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.904724 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.904808 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt8kk\" (UniqueName: \"kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.904845 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.904867 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.904931 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.905030 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.905073 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.905163 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data\") pod \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\" (UID: \"0c60c27d-e549-4ced-a2eb-6e400cb110cc\") " Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.905373 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.905707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.906265 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.906289 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c60c27d-e549-4ced-a2eb-6e400cb110cc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.910154 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk" (OuterVolumeSpecName: "kube-api-access-kt8kk") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "kube-api-access-kt8kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.910679 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts" (OuterVolumeSpecName: "scripts") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.936823 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.969442 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:40 crc kubenswrapper[4989]: I1006 09:03:40.989888 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.007541 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.007575 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt8kk\" (UniqueName: \"kubernetes.io/projected/0c60c27d-e549-4ced-a2eb-6e400cb110cc-kube-api-access-kt8kk\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.007588 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.007600 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.007611 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.043008 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data" (OuterVolumeSpecName: "config-data") pod "0c60c27d-e549-4ced-a2eb-6e400cb110cc" (UID: "0c60c27d-e549-4ced-a2eb-6e400cb110cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.109375 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c60c27d-e549-4ced-a2eb-6e400cb110cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.301607 4989 generic.go:334] "Generic (PLEG): container finished" podID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerID="901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30" exitCode=0 Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.301676 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerDied","Data":"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30"} Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.301747 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c60c27d-e549-4ced-a2eb-6e400cb110cc","Type":"ContainerDied","Data":"9c4a64bb2ee45c6cf2ee200ac852959d7502554f9a2c1442f4536087ac8d14fe"} Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.301771 4989 scope.go:117] "RemoveContainer" containerID="0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.301796 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.384822 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.387786 4989 scope.go:117] "RemoveContainer" containerID="e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.397056 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.412766 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413257 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="registry-server" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413274 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="registry-server" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413294 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="extract-content" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413301 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="extract-content" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413317 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="sg-core" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413323 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="sg-core" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413338 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="extract-utilities" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413347 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="extract-utilities" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413357 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-central-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413363 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-central-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413373 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="proxy-httpd" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413379 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="proxy-httpd" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.413395 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-notification-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413401 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-notification-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413568 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="proxy-httpd" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413580 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="sg-core" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413594 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-notification-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413600 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" containerName="ceilometer-central-agent" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.413615 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ddc1aa3-4dd2-4208-a199-2b867a42968f" containerName="registry-server" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.418705 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.421330 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.421359 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.421619 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.442060 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.452544 4989 scope.go:117] "RemoveContainer" containerID="901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.514919 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.514962 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515014 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515033 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515146 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515225 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.515248 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp84m\" (UniqueName: \"kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.558686 4989 scope.go:117] "RemoveContainer" containerID="182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.589272 4989 scope.go:117] "RemoveContainer" containerID="0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.590481 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07\": container with ID starting with 0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07 not found: ID does not exist" containerID="0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.590512 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07"} err="failed to get container status \"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07\": rpc error: code = NotFound desc = could not find container \"0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07\": container with ID starting with 0d4e694de4d4d839f8bd5c1cd48da06f3180b4685897e9ab71f4d6de6f531a07 not found: ID does not exist" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.590532 4989 scope.go:117] "RemoveContainer" containerID="e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.591793 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208\": container with ID starting with e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208 not found: ID does not exist" containerID="e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.591817 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208"} err="failed to get container status \"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208\": rpc error: code = NotFound desc = could not find container \"e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208\": container with ID starting with e8d6573bca4ca50f1132e8747115c6d5225286121af11ff917bae9b394505208 not found: ID does not exist" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.591830 4989 scope.go:117] "RemoveContainer" containerID="901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.592198 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30\": container with ID starting with 901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30 not found: ID does not exist" containerID="901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.592307 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30"} err="failed to get container status \"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30\": rpc error: code = NotFound desc = could not find container \"901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30\": container with ID starting with 901c91da19e2ae504d54a632fced92ff78a9bd9cdef63acaeb25bc7e37d83a30 not found: ID does not exist" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.592340 4989 scope.go:117] "RemoveContainer" containerID="182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d" Oct 06 09:03:41 crc kubenswrapper[4989]: E1006 09:03:41.592776 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d\": container with ID starting with 182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d not found: ID does not exist" containerID="182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.592804 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d"} err="failed to get container status \"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d\": rpc error: code = NotFound desc = could not find container \"182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d\": container with ID starting with 182d64cfd9dcbf922d2266cf03e963de3d3dafc9055bfa83fb44f2051ff5377d not found: ID does not exist" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.600694 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617409 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp84m\" (UniqueName: \"kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617471 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617590 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617621 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.617817 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.618735 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.619212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.623640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.624543 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.624629 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.626223 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.628763 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.637974 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp84m\" (UniqueName: \"kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m\") pod \"ceilometer-0\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.660098 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.853600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.858821 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:41 crc kubenswrapper[4989]: I1006 09:03:41.972309 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c60c27d-e549-4ced-a2eb-6e400cb110cc" path="/var/lib/kubelet/pods/0c60c27d-e549-4ced-a2eb-6e400cb110cc/volumes" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.027674 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle\") pod \"461d00bd-1c96-4406-a64a-32f159af8a1b\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.027768 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs\") pod \"461d00bd-1c96-4406-a64a-32f159af8a1b\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.027829 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data\") pod \"461d00bd-1c96-4406-a64a-32f159af8a1b\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.027862 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r98zv\" (UniqueName: \"kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv\") pod \"461d00bd-1c96-4406-a64a-32f159af8a1b\" (UID: \"461d00bd-1c96-4406-a64a-32f159af8a1b\") " Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.028626 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs" (OuterVolumeSpecName: "logs") pod "461d00bd-1c96-4406-a64a-32f159af8a1b" (UID: "461d00bd-1c96-4406-a64a-32f159af8a1b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.036114 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv" (OuterVolumeSpecName: "kube-api-access-r98zv") pod "461d00bd-1c96-4406-a64a-32f159af8a1b" (UID: "461d00bd-1c96-4406-a64a-32f159af8a1b"). InnerVolumeSpecName "kube-api-access-r98zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.056718 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "461d00bd-1c96-4406-a64a-32f159af8a1b" (UID: "461d00bd-1c96-4406-a64a-32f159af8a1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.056774 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data" (OuterVolumeSpecName: "config-data") pod "461d00bd-1c96-4406-a64a-32f159af8a1b" (UID: "461d00bd-1c96-4406-a64a-32f159af8a1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.129596 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d00bd-1c96-4406-a64a-32f159af8a1b-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.129636 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.129648 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r98zv\" (UniqueName: \"kubernetes.io/projected/461d00bd-1c96-4406-a64a-32f159af8a1b-kube-api-access-r98zv\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.129675 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d00bd-1c96-4406-a64a-32f159af8a1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.309990 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.324203 4989 generic.go:334] "Generic (PLEG): container finished" podID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerID="88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d" exitCode=0 Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.325824 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerDied","Data":"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d"} Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.325863 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.325925 4989 scope.go:117] "RemoveContainer" containerID="88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.325907 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"461d00bd-1c96-4406-a64a-32f159af8a1b","Type":"ContainerDied","Data":"0ef058e4e993d253d6f0f0af691f7a349d9891b16af050a3183a6fa4b0a278d7"} Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.346031 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.453037 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.463777 4989 scope.go:117] "RemoveContainer" containerID="8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.469871 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.493703 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:42 crc kubenswrapper[4989]: E1006 09:03:42.494128 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-log" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.494146 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-log" Oct 06 09:03:42 crc kubenswrapper[4989]: E1006 09:03:42.494168 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-api" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.494177 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-api" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.494356 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-api" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.494370 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" containerName="nova-api-log" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.494431 4989 scope.go:117] "RemoveContainer" containerID="88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d" Oct 06 09:03:42 crc kubenswrapper[4989]: E1006 09:03:42.495046 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d\": container with ID starting with 88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d not found: ID does not exist" containerID="88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.495093 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d"} err="failed to get container status \"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d\": rpc error: code = NotFound desc = could not find container \"88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d\": container with ID starting with 88ee14fc8481219c0119d4098b84aa60dde8dd478b6d39d993f2723acba6ec4d not found: ID does not exist" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.495118 4989 scope.go:117] "RemoveContainer" containerID="8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.495366 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: E1006 09:03:42.495631 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4\": container with ID starting with 8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4 not found: ID does not exist" containerID="8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.495858 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4"} err="failed to get container status \"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4\": rpc error: code = NotFound desc = could not find container \"8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4\": container with ID starting with 8d23b057e664695c8508d91ab86fea2ce657809947725a9eef972665f1392bc4 not found: ID does not exist" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.502329 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.502591 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.502843 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.510346 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.604396 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-xdjg5"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.606108 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.611769 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.612394 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.614871 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xdjg5"] Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639342 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7jtd\" (UniqueName: \"kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639469 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639515 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639545 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.639593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741584 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741630 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741722 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741750 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741771 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741793 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741808 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741865 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741882 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhjb\" (UniqueName: \"kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.741907 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7jtd\" (UniqueName: \"kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.742792 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.747345 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.747922 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.748283 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.751692 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.762596 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7jtd\" (UniqueName: \"kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd\") pod \"nova-api-0\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.823107 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.848329 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.848431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.848525 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.848549 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhjb\" (UniqueName: \"kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.851703 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.851711 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.852583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.866901 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhjb\" (UniqueName: \"kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb\") pod \"nova-cell1-cell-mapping-xdjg5\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:42 crc kubenswrapper[4989]: I1006 09:03:42.922835 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:43 crc kubenswrapper[4989]: I1006 09:03:43.262752 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xdjg5"] Oct 06 09:03:43 crc kubenswrapper[4989]: I1006 09:03:43.348002 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xdjg5" event={"ID":"6d9edea7-f372-42ff-bd54-bb1a919f79ac","Type":"ContainerStarted","Data":"43e166bc0fd2f5ca9fde7a6f60344e881654c0612aa42dcdca87444e75285775"} Oct 06 09:03:43 crc kubenswrapper[4989]: I1006 09:03:43.349017 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:43 crc kubenswrapper[4989]: I1006 09:03:43.362383 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerStarted","Data":"12b835224b464f10981293d881a362bbc1d3a5512530f5acba148e8ca14df5e3"} Oct 06 09:03:43 crc kubenswrapper[4989]: I1006 09:03:43.959965 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="461d00bd-1c96-4406-a64a-32f159af8a1b" path="/var/lib/kubelet/pods/461d00bd-1c96-4406-a64a-32f159af8a1b/volumes" Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.405498 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerStarted","Data":"0e944028fd090111d4e8a713c7330ed17face3e9d21c00ba187b11ee5fe64bf4"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.405568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerStarted","Data":"1113ed9589562e5b58b800ea0d696a3d844f0b50b70b74e3d2b8d7f5c23c5eb4"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.408250 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerStarted","Data":"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.408289 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerStarted","Data":"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.408302 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerStarted","Data":"7231a591399e91a613d146c6e48b7154508c6b9e6ddea760e62d1727fe145335"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.413183 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xdjg5" event={"ID":"6d9edea7-f372-42ff-bd54-bb1a919f79ac","Type":"ContainerStarted","Data":"385a5a4463655d9562fa75c66db444ad5a9ac4c63973728715660d99561581d1"} Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.462855 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.462833791 podStartE2EDuration="2.462833791s" podCreationTimestamp="2025-10-06 09:03:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:44.433512895 +0000 UTC m=+1475.223538465" watchObservedRunningTime="2025-10-06 09:03:44.462833791 +0000 UTC m=+1475.252859381" Oct 06 09:03:44 crc kubenswrapper[4989]: I1006 09:03:44.465363 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-xdjg5" podStartSLOduration=2.465354333 podStartE2EDuration="2.465354333s" podCreationTimestamp="2025-10-06 09:03:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:44.453209193 +0000 UTC m=+1475.243234773" watchObservedRunningTime="2025-10-06 09:03:44.465354333 +0000 UTC m=+1475.255379933" Oct 06 09:03:45 crc kubenswrapper[4989]: I1006 09:03:45.424861 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerStarted","Data":"6a8db167aeebd3ccb93770c425902d5284cd7da3659d4cafff002f345363260f"} Oct 06 09:03:45 crc kubenswrapper[4989]: I1006 09:03:45.743617 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:03:45 crc kubenswrapper[4989]: I1006 09:03:45.799264 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:03:45 crc kubenswrapper[4989]: I1006 09:03:45.799723 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="dnsmasq-dns" containerID="cri-o://000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba" gracePeriod=10 Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.367929 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446325 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446505 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446539 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446632 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446698 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.446739 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp48m\" (UniqueName: \"kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m\") pod \"8459d522-eb54-4580-b3ca-39e54c86a3c5\" (UID: \"8459d522-eb54-4580-b3ca-39e54c86a3c5\") " Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.449928 4989 generic.go:334] "Generic (PLEG): container finished" podID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerID="000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba" exitCode=0 Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.449970 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" event={"ID":"8459d522-eb54-4580-b3ca-39e54c86a3c5","Type":"ContainerDied","Data":"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba"} Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.449999 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" event={"ID":"8459d522-eb54-4580-b3ca-39e54c86a3c5","Type":"ContainerDied","Data":"a541047321fd3b6cc60b201f97e426d45c82cb179cff3f28a78ef86d776e9134"} Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.450031 4989 scope.go:117] "RemoveContainer" containerID="000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.450121 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-chph9" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.461131 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m" (OuterVolumeSpecName: "kube-api-access-vp48m") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "kube-api-access-vp48m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.524822 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.550179 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.550216 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp48m\" (UniqueName: \"kubernetes.io/projected/8459d522-eb54-4580-b3ca-39e54c86a3c5-kube-api-access-vp48m\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.552030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.552942 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.556428 4989 scope.go:117] "RemoveContainer" containerID="0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.568174 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config" (OuterVolumeSpecName: "config") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.572112 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8459d522-eb54-4580-b3ca-39e54c86a3c5" (UID: "8459d522-eb54-4580-b3ca-39e54c86a3c5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.589802 4989 scope.go:117] "RemoveContainer" containerID="000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba" Oct 06 09:03:46 crc kubenswrapper[4989]: E1006 09:03:46.590233 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba\": container with ID starting with 000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba not found: ID does not exist" containerID="000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.590272 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba"} err="failed to get container status \"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba\": rpc error: code = NotFound desc = could not find container \"000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba\": container with ID starting with 000fd2d48e37022cc3c0261831f11f5f3bf67143544d7c7c99b78926ccc49fba not found: ID does not exist" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.590292 4989 scope.go:117] "RemoveContainer" containerID="0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a" Oct 06 09:03:46 crc kubenswrapper[4989]: E1006 09:03:46.592338 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a\": container with ID starting with 0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a not found: ID does not exist" containerID="0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.592363 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a"} err="failed to get container status \"0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a\": rpc error: code = NotFound desc = could not find container \"0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a\": container with ID starting with 0f62c0770cc4a53c930305b1fdf7d9d3bff939d2f580d88039124f590f06de8a not found: ID does not exist" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.654870 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.654912 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.654922 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.654931 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8459d522-eb54-4580-b3ca-39e54c86a3c5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.790670 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:03:46 crc kubenswrapper[4989]: I1006 09:03:46.799534 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-chph9"] Oct 06 09:03:47 crc kubenswrapper[4989]: I1006 09:03:47.471642 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerStarted","Data":"a3e857f4d02909a9a5deea6f9f5ec9a459f3932e69093005ea9d3a9d28adff43"} Oct 06 09:03:47 crc kubenswrapper[4989]: I1006 09:03:47.473060 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 09:03:47 crc kubenswrapper[4989]: I1006 09:03:47.506492 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.448404924 podStartE2EDuration="6.50646684s" podCreationTimestamp="2025-10-06 09:03:41 +0000 UTC" firstStartedPulling="2025-10-06 09:03:42.315867751 +0000 UTC m=+1473.105893331" lastFinishedPulling="2025-10-06 09:03:46.373929667 +0000 UTC m=+1477.163955247" observedRunningTime="2025-10-06 09:03:47.498367326 +0000 UTC m=+1478.288392916" watchObservedRunningTime="2025-10-06 09:03:47.50646684 +0000 UTC m=+1478.296492420" Oct 06 09:03:47 crc kubenswrapper[4989]: I1006 09:03:47.946992 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" path="/var/lib/kubelet/pods/8459d522-eb54-4580-b3ca-39e54c86a3c5/volumes" Oct 06 09:03:49 crc kubenswrapper[4989]: I1006 09:03:49.497403 4989 generic.go:334] "Generic (PLEG): container finished" podID="6d9edea7-f372-42ff-bd54-bb1a919f79ac" containerID="385a5a4463655d9562fa75c66db444ad5a9ac4c63973728715660d99561581d1" exitCode=0 Oct 06 09:03:49 crc kubenswrapper[4989]: I1006 09:03:49.497517 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xdjg5" event={"ID":"6d9edea7-f372-42ff-bd54-bb1a919f79ac","Type":"ContainerDied","Data":"385a5a4463655d9562fa75c66db444ad5a9ac4c63973728715660d99561581d1"} Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.859720 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.938687 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdhjb\" (UniqueName: \"kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb\") pod \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.938757 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data\") pod \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.938878 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle\") pod \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.938968 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts\") pod \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\" (UID: \"6d9edea7-f372-42ff-bd54-bb1a919f79ac\") " Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.945155 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts" (OuterVolumeSpecName: "scripts") pod "6d9edea7-f372-42ff-bd54-bb1a919f79ac" (UID: "6d9edea7-f372-42ff-bd54-bb1a919f79ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.945460 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb" (OuterVolumeSpecName: "kube-api-access-pdhjb") pod "6d9edea7-f372-42ff-bd54-bb1a919f79ac" (UID: "6d9edea7-f372-42ff-bd54-bb1a919f79ac"). InnerVolumeSpecName "kube-api-access-pdhjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.966722 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data" (OuterVolumeSpecName: "config-data") pod "6d9edea7-f372-42ff-bd54-bb1a919f79ac" (UID: "6d9edea7-f372-42ff-bd54-bb1a919f79ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:50.970908 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d9edea7-f372-42ff-bd54-bb1a919f79ac" (UID: "6d9edea7-f372-42ff-bd54-bb1a919f79ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.041494 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdhjb\" (UniqueName: \"kubernetes.io/projected/6d9edea7-f372-42ff-bd54-bb1a919f79ac-kube-api-access-pdhjb\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.041525 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.041534 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.041543 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d9edea7-f372-42ff-bd54-bb1a919f79ac-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.518696 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xdjg5" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.518630 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xdjg5" event={"ID":"6d9edea7-f372-42ff-bd54-bb1a919f79ac","Type":"ContainerDied","Data":"43e166bc0fd2f5ca9fde7a6f60344e881654c0612aa42dcdca87444e75285775"} Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.518746 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43e166bc0fd2f5ca9fde7a6f60344e881654c0612aa42dcdca87444e75285775" Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.697272 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.697555 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-log" containerID="cri-o://293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" gracePeriod=30 Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.697641 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-api" containerID="cri-o://4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" gracePeriod=30 Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.736538 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.737068 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerName="nova-scheduler-scheduler" containerID="cri-o://d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" gracePeriod=30 Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.749991 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.750232 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" containerID="cri-o://76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5" gracePeriod=30 Oct 06 09:03:51 crc kubenswrapper[4989]: I1006 09:03:51.750329 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" containerID="cri-o://b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3" gracePeriod=30 Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.251613 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370096 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370157 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370300 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7jtd\" (UniqueName: \"kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370438 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370517 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs\") pod \"f58cd58a-024d-4aea-ab69-59c68e587785\" (UID: \"f58cd58a-024d-4aea-ab69-59c68e587785\") " Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.370741 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs" (OuterVolumeSpecName: "logs") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.371281 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f58cd58a-024d-4aea-ab69-59c68e587785-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.380895 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd" (OuterVolumeSpecName: "kube-api-access-s7jtd") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "kube-api-access-s7jtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.398873 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.404981 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data" (OuterVolumeSpecName: "config-data") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.407083 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.408549 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.410054 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.410092 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerName="nova-scheduler-scheduler" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.425935 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.439133 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f58cd58a-024d-4aea-ab69-59c68e587785" (UID: "f58cd58a-024d-4aea-ab69-59c68e587785"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.473497 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7jtd\" (UniqueName: \"kubernetes.io/projected/f58cd58a-024d-4aea-ab69-59c68e587785-kube-api-access-s7jtd\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.473546 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.473558 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.473568 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.473579 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f58cd58a-024d-4aea-ab69-59c68e587785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.528436 4989 generic.go:334] "Generic (PLEG): container finished" podID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerID="76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5" exitCode=143 Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.528620 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerDied","Data":"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5"} Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530683 4989 generic.go:334] "Generic (PLEG): container finished" podID="f58cd58a-024d-4aea-ab69-59c68e587785" containerID="4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" exitCode=0 Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530704 4989 generic.go:334] "Generic (PLEG): container finished" podID="f58cd58a-024d-4aea-ab69-59c68e587785" containerID="293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" exitCode=143 Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530721 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerDied","Data":"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e"} Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530737 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerDied","Data":"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb"} Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f58cd58a-024d-4aea-ab69-59c68e587785","Type":"ContainerDied","Data":"7231a591399e91a613d146c6e48b7154508c6b9e6ddea760e62d1727fe145335"} Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530762 4989 scope.go:117] "RemoveContainer" containerID="4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.530879 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.563017 4989 scope.go:117] "RemoveContainer" containerID="293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.570580 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.590254 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.600789 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.601197 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="dnsmasq-dns" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601209 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="dnsmasq-dns" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.601228 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-log" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601234 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-log" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.601244 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-api" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601250 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-api" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.601272 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="init" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601278 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="init" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.601301 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9edea7-f372-42ff-bd54-bb1a919f79ac" containerName="nova-manage" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601306 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9edea7-f372-42ff-bd54-bb1a919f79ac" containerName="nova-manage" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601480 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8459d522-eb54-4580-b3ca-39e54c86a3c5" containerName="dnsmasq-dns" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601498 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-api" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601507 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9edea7-f372-42ff-bd54-bb1a919f79ac" containerName="nova-manage" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.601517 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" containerName="nova-api-log" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.602646 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.608816 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.609545 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.609566 4989 scope.go:117] "RemoveContainer" containerID="4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.609730 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.609839 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.612018 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e\": container with ID starting with 4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e not found: ID does not exist" containerID="4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612054 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e"} err="failed to get container status \"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e\": rpc error: code = NotFound desc = could not find container \"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e\": container with ID starting with 4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e not found: ID does not exist" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612076 4989 scope.go:117] "RemoveContainer" containerID="293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" Oct 06 09:03:52 crc kubenswrapper[4989]: E1006 09:03:52.612524 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb\": container with ID starting with 293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb not found: ID does not exist" containerID="293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612540 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb"} err="failed to get container status \"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb\": rpc error: code = NotFound desc = could not find container \"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb\": container with ID starting with 293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb not found: ID does not exist" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612552 4989 scope.go:117] "RemoveContainer" containerID="4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612875 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e"} err="failed to get container status \"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e\": rpc error: code = NotFound desc = could not find container \"4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e\": container with ID starting with 4623fdaa70e8e2c0e7a8272bd86f6f52bd79a5591996d2dd83191c792290506e not found: ID does not exist" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.612890 4989 scope.go:117] "RemoveContainer" containerID="293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.613976 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb"} err="failed to get container status \"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb\": rpc error: code = NotFound desc = could not find container \"293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb\": container with ID starting with 293524cbce1b8806f6d1aea5fda78fa94b137bd102e99d0b93a9bd13f70643cb not found: ID does not exist" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677691 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677740 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94gp5\" (UniqueName: \"kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677784 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677809 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677906 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.677921 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.794967 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.795029 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94gp5\" (UniqueName: \"kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.795127 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.795173 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.795272 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.795300 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.796191 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.800014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.801426 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.801630 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.801852 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.815323 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94gp5\" (UniqueName: \"kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5\") pod \"nova-api-0\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " pod="openstack/nova-api-0" Oct 06 09:03:52 crc kubenswrapper[4989]: I1006 09:03:52.929958 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:03:53 crc kubenswrapper[4989]: I1006 09:03:53.357103 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:03:53 crc kubenswrapper[4989]: I1006 09:03:53.540111 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerStarted","Data":"9e2d55d5310b2a032c3053d2f611de15d5d3b47bc88f9275f2f82580ad17ce65"} Oct 06 09:03:53 crc kubenswrapper[4989]: I1006 09:03:53.540992 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerStarted","Data":"2c549db93f26ee1b5fe2037813724772ee6b80e80fdbb7d7333383eba9d8bab2"} Oct 06 09:03:53 crc kubenswrapper[4989]: I1006 09:03:53.949834 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f58cd58a-024d-4aea-ab69-59c68e587785" path="/var/lib/kubelet/pods/f58cd58a-024d-4aea-ab69-59c68e587785/volumes" Oct 06 09:03:54 crc kubenswrapper[4989]: I1006 09:03:54.554385 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerStarted","Data":"de97aa68267d73cae2bc02e80acd84df5900e6e32436971bd32cc722e73765e7"} Oct 06 09:03:54 crc kubenswrapper[4989]: I1006 09:03:54.576058 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.576038179 podStartE2EDuration="2.576038179s" podCreationTimestamp="2025-10-06 09:03:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:54.569883861 +0000 UTC m=+1485.359909451" watchObservedRunningTime="2025-10-06 09:03:54.576038179 +0000 UTC m=+1485.366063759" Oct 06 09:03:54 crc kubenswrapper[4989]: I1006 09:03:54.873197 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:34830->10.217.0.189:8775: read: connection reset by peer" Oct 06 09:03:54 crc kubenswrapper[4989]: I1006 09:03:54.873256 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:34846->10.217.0.189:8775: read: connection reset by peer" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.309087 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.446572 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs\") pod \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.446756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle\") pod \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.446784 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvs2d\" (UniqueName: \"kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d\") pod \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.446819 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs\") pod \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.446839 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data\") pod \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\" (UID: \"71de8c65-18ee-4eb8-a461-f71dd3db00ce\") " Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.447557 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs" (OuterVolumeSpecName: "logs") pod "71de8c65-18ee-4eb8-a461-f71dd3db00ce" (UID: "71de8c65-18ee-4eb8-a461-f71dd3db00ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.451861 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d" (OuterVolumeSpecName: "kube-api-access-wvs2d") pod "71de8c65-18ee-4eb8-a461-f71dd3db00ce" (UID: "71de8c65-18ee-4eb8-a461-f71dd3db00ce"). InnerVolumeSpecName "kube-api-access-wvs2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.473228 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71de8c65-18ee-4eb8-a461-f71dd3db00ce" (UID: "71de8c65-18ee-4eb8-a461-f71dd3db00ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.478642 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data" (OuterVolumeSpecName: "config-data") pod "71de8c65-18ee-4eb8-a461-f71dd3db00ce" (UID: "71de8c65-18ee-4eb8-a461-f71dd3db00ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.503758 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "71de8c65-18ee-4eb8-a461-f71dd3db00ce" (UID: "71de8c65-18ee-4eb8-a461-f71dd3db00ce"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.548840 4989 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.548882 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.548895 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvs2d\" (UniqueName: \"kubernetes.io/projected/71de8c65-18ee-4eb8-a461-f71dd3db00ce-kube-api-access-wvs2d\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.548908 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71de8c65-18ee-4eb8-a461-f71dd3db00ce-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.548921 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71de8c65-18ee-4eb8-a461-f71dd3db00ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.564053 4989 generic.go:334] "Generic (PLEG): container finished" podID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerID="b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3" exitCode=0 Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.564083 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.564151 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerDied","Data":"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3"} Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.564178 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"71de8c65-18ee-4eb8-a461-f71dd3db00ce","Type":"ContainerDied","Data":"0897d1eeeb60ab92ac6614f1dc3ab345644a6c768024308f83211bd634387ed0"} Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.564195 4989 scope.go:117] "RemoveContainer" containerID="b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.589882 4989 scope.go:117] "RemoveContainer" containerID="76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.606646 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.619042 4989 scope.go:117] "RemoveContainer" containerID="b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3" Oct 06 09:03:55 crc kubenswrapper[4989]: E1006 09:03:55.619530 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3\": container with ID starting with b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3 not found: ID does not exist" containerID="b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.619586 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3"} err="failed to get container status \"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3\": rpc error: code = NotFound desc = could not find container \"b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3\": container with ID starting with b3150cdfceab392b9f3dbec894e3a2458150dc8043925e98e6f83c5ea37489e3 not found: ID does not exist" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.619618 4989 scope.go:117] "RemoveContainer" containerID="76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5" Oct 06 09:03:55 crc kubenswrapper[4989]: E1006 09:03:55.620081 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5\": container with ID starting with 76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5 not found: ID does not exist" containerID="76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.620160 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5"} err="failed to get container status \"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5\": rpc error: code = NotFound desc = could not find container \"76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5\": container with ID starting with 76035f1dc484a1da66417e3977b8f0e749da7560f467db272f2dbfee50a80db5 not found: ID does not exist" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.622396 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.632341 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:55 crc kubenswrapper[4989]: E1006 09:03:55.632892 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.632914 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" Oct 06 09:03:55 crc kubenswrapper[4989]: E1006 09:03:55.632939 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.632946 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.633139 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-log" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.633157 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" containerName="nova-metadata-metadata" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.634165 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.641563 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.642297 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.642472 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.752962 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.753869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.753968 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.754061 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.754139 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.855802 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.855870 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.855889 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.855910 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.855973 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.856419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.859892 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.860461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.861309 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.872038 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp\") pod \"nova-metadata-0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " pod="openstack/nova-metadata-0" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.950877 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71de8c65-18ee-4eb8-a461-f71dd3db00ce" path="/var/lib/kubelet/pods/71de8c65-18ee-4eb8-a461-f71dd3db00ce/volumes" Oct 06 09:03:55 crc kubenswrapper[4989]: I1006 09:03:55.958427 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.390979 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.537747 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.581362 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerStarted","Data":"648b0a633bb695eac1b4b106abd3e4423a4c45eceef2c1366e8d7b5a4a2862d4"} Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.581460 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerStarted","Data":"ae61dbab66d5f5da263e5dfec201cabd998f00bb8e59729d685cf7a7e5ffb8ee"} Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.584118 4989 generic.go:334] "Generic (PLEG): container finished" podID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" exitCode=0 Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.584165 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aeb4f1b9-106e-42a4-93b1-292448b7e557","Type":"ContainerDied","Data":"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0"} Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.584198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aeb4f1b9-106e-42a4-93b1-292448b7e557","Type":"ContainerDied","Data":"f649817a9f7ee0ca05c7720d837459a76ea2d0868546d3044e5d87c4420f4bb4"} Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.584196 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.584220 4989 scope.go:117] "RemoveContainer" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.631604 4989 scope.go:117] "RemoveContainer" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" Oct 06 09:03:56 crc kubenswrapper[4989]: E1006 09:03:56.632488 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0\": container with ID starting with d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0 not found: ID does not exist" containerID="d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.632564 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0"} err="failed to get container status \"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0\": rpc error: code = NotFound desc = could not find container \"d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0\": container with ID starting with d33bd1c32c1ddc8b3cfd1caecb5740afcadbea2b01b783e449efc813214e80e0 not found: ID does not exist" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.676193 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data\") pod \"aeb4f1b9-106e-42a4-93b1-292448b7e557\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.676350 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc9kh\" (UniqueName: \"kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh\") pod \"aeb4f1b9-106e-42a4-93b1-292448b7e557\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.676616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle\") pod \"aeb4f1b9-106e-42a4-93b1-292448b7e557\" (UID: \"aeb4f1b9-106e-42a4-93b1-292448b7e557\") " Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.681969 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh" (OuterVolumeSpecName: "kube-api-access-rc9kh") pod "aeb4f1b9-106e-42a4-93b1-292448b7e557" (UID: "aeb4f1b9-106e-42a4-93b1-292448b7e557"). InnerVolumeSpecName "kube-api-access-rc9kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.712483 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aeb4f1b9-106e-42a4-93b1-292448b7e557" (UID: "aeb4f1b9-106e-42a4-93b1-292448b7e557"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.717727 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data" (OuterVolumeSpecName: "config-data") pod "aeb4f1b9-106e-42a4-93b1-292448b7e557" (UID: "aeb4f1b9-106e-42a4-93b1-292448b7e557"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.779702 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.779742 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc9kh\" (UniqueName: \"kubernetes.io/projected/aeb4f1b9-106e-42a4-93b1-292448b7e557-kube-api-access-rc9kh\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.779759 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeb4f1b9-106e-42a4-93b1-292448b7e557-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.931883 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.956311 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.982729 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:56 crc kubenswrapper[4989]: E1006 09:03:56.983188 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerName="nova-scheduler-scheduler" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.983212 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerName="nova-scheduler-scheduler" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.983442 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" containerName="nova-scheduler-scheduler" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.984122 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.990883 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:56 crc kubenswrapper[4989]: I1006 09:03:56.994038 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.084594 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.084680 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.084966 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcc9x\" (UniqueName: \"kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.187151 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.187241 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.187331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcc9x\" (UniqueName: \"kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.191300 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.193721 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.206014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcc9x\" (UniqueName: \"kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x\") pod \"nova-scheduler-0\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.347102 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.595634 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerStarted","Data":"41c3a3ea88eda96a1475512601b4880062d4b09816817a60f531e6bc6a058912"} Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.618435 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6184180120000002 podStartE2EDuration="2.618418012s" podCreationTimestamp="2025-10-06 09:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:57.616794565 +0000 UTC m=+1488.406820155" watchObservedRunningTime="2025-10-06 09:03:57.618418012 +0000 UTC m=+1488.408443592" Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.786677 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:03:57 crc kubenswrapper[4989]: W1006 09:03:57.794110 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod716f391a_794d_4511_9fc2_fa48a3a948cc.slice/crio-04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f WatchSource:0}: Error finding container 04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f: Status 404 returned error can't find the container with id 04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f Oct 06 09:03:57 crc kubenswrapper[4989]: I1006 09:03:57.951543 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb4f1b9-106e-42a4-93b1-292448b7e557" path="/var/lib/kubelet/pods/aeb4f1b9-106e-42a4-93b1-292448b7e557/volumes" Oct 06 09:03:58 crc kubenswrapper[4989]: I1006 09:03:58.613568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"716f391a-794d-4511-9fc2-fa48a3a948cc","Type":"ContainerStarted","Data":"051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774"} Oct 06 09:03:58 crc kubenswrapper[4989]: I1006 09:03:58.613946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"716f391a-794d-4511-9fc2-fa48a3a948cc","Type":"ContainerStarted","Data":"04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f"} Oct 06 09:03:58 crc kubenswrapper[4989]: I1006 09:03:58.644146 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.644120134 podStartE2EDuration="2.644120134s" podCreationTimestamp="2025-10-06 09:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:03:58.632178279 +0000 UTC m=+1489.422203899" watchObservedRunningTime="2025-10-06 09:03:58.644120134 +0000 UTC m=+1489.434145744" Oct 06 09:04:00 crc kubenswrapper[4989]: I1006 09:04:00.960197 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:04:00 crc kubenswrapper[4989]: I1006 09:04:00.960546 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 09:04:02 crc kubenswrapper[4989]: I1006 09:04:02.347411 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 09:04:02 crc kubenswrapper[4989]: I1006 09:04:02.930912 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:04:02 crc kubenswrapper[4989]: I1006 09:04:02.931022 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.948862 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.948929 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.949254 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.949296 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.950483 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.951408 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:04:03 crc kubenswrapper[4989]: I1006 09:04:03.951502 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec" gracePeriod=600 Oct 06 09:04:04 crc kubenswrapper[4989]: I1006 09:04:04.680265 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec"} Oct 06 09:04:04 crc kubenswrapper[4989]: I1006 09:04:04.680566 4989 scope.go:117] "RemoveContainer" containerID="c20a21fd864ec7ce9c4b090aadfbfbc35905aa97fe48fb26ac2017ecf202d050" Oct 06 09:04:04 crc kubenswrapper[4989]: I1006 09:04:04.680264 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec" exitCode=0 Oct 06 09:04:04 crc kubenswrapper[4989]: I1006 09:04:04.680628 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6"} Oct 06 09:04:05 crc kubenswrapper[4989]: I1006 09:04:05.959717 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 09:04:05 crc kubenswrapper[4989]: I1006 09:04:05.960095 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 09:04:06 crc kubenswrapper[4989]: I1006 09:04:06.973818 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:06 crc kubenswrapper[4989]: I1006 09:04:06.973871 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:07 crc kubenswrapper[4989]: I1006 09:04:07.348172 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 09:04:07 crc kubenswrapper[4989]: I1006 09:04:07.380336 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 09:04:07 crc kubenswrapper[4989]: I1006 09:04:07.743442 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 09:04:11 crc kubenswrapper[4989]: I1006 09:04:11.862459 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 09:04:12 crc kubenswrapper[4989]: I1006 09:04:12.937420 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 09:04:12 crc kubenswrapper[4989]: I1006 09:04:12.938051 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 09:04:12 crc kubenswrapper[4989]: I1006 09:04:12.948530 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 09:04:12 crc kubenswrapper[4989]: I1006 09:04:12.951260 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 09:04:13 crc kubenswrapper[4989]: I1006 09:04:13.764445 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 09:04:13 crc kubenswrapper[4989]: I1006 09:04:13.770893 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 09:04:15 crc kubenswrapper[4989]: I1006 09:04:15.966522 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 09:04:15 crc kubenswrapper[4989]: I1006 09:04:15.967029 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 09:04:15 crc kubenswrapper[4989]: I1006 09:04:15.974157 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 09:04:15 crc kubenswrapper[4989]: I1006 09:04:15.974722 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.415336 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.422671 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" containerName="openstackclient" containerID="cri-o://ca2377b7cd7f73d36c389bf39d5a5b793d4492f8b024e80813557074b081fb4c" gracePeriod=2 Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.444781 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.477685 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.477978 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="cinder-scheduler" containerID="cri-o://f4b25eb0db619fd2778bd141569f6d60f1ea4b0b8e5c38cb07786cb6a462723c" gracePeriod=30 Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.478137 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="probe" containerID="cri-o://e2d916bd8ce663a1374a2102f14126cf7f39b34e74293d78b95861c7f07ea483" gracePeriod=30 Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.589723 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:04:34 crc kubenswrapper[4989]: E1006 09:04:34.590189 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" containerName="openstackclient" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.590202 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" containerName="openstackclient" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.590359 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" containerName="openstackclient" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.591277 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.602946 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.605020 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.652812 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.681849 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.692226 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720265 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720352 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssmsk\" (UniqueName: \"kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720391 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720456 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720488 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720531 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fmfm\" (UniqueName: \"kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720588 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720667 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.720685 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.760395 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.760845 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api-log" containerID="cri-o://61197413d388b49b459e9e0efe53d5c352fb7b325ef677d33c7410cb203309dc" gracePeriod=30 Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.761248 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api" containerID="cri-o://0bf97fec682ff1dbd5c988be1156a9e788940eb34663d369552ca96c1e405f14" gracePeriod=30 Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822030 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fmfm\" (UniqueName: \"kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822142 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822213 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822241 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822281 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822307 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssmsk\" (UniqueName: \"kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822415 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.822546 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.828300 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.831241 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: E1006 09:04:34.831309 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 06 09:04:34 crc kubenswrapper[4989]: E1006 09:04:34.831383 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data podName:03d954db-7dc6-4921-b260-1c189b9492c2 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:35.331363389 +0000 UTC m=+1526.121388969 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data") pod "rabbitmq-server-0" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2") : configmap "rabbitmq-config-data" not found Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.843722 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.848737 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.848923 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.851529 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.861208 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.877393 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.878910 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.887759 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.888868 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fmfm\" (UniqueName: \"kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm\") pod \"barbican-keystone-listener-567987cf6d-htd97\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.903233 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssmsk\" (UniqueName: \"kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.904360 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle\") pod \"barbican-worker-68c46dd79-4twxt\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.918311 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:04:34 crc kubenswrapper[4989]: I1006 09:04:34.965175 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040245 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgx8r\" (UniqueName: \"kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040511 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040574 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040610 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040629 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.040722 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.055810 4989 generic.go:334] "Generic (PLEG): container finished" podID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerID="61197413d388b49b459e9e0efe53d5c352fb7b325ef677d33c7410cb203309dc" exitCode=143 Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.056032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerDied","Data":"61197413d388b49b459e9e0efe53d5c352fb7b325ef677d33c7410cb203309dc"} Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.116395 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143221 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143496 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143593 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143677 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143800 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143900 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.143970 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgx8r\" (UniqueName: \"kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.153332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.164614 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.181455 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.181545 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.182122 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="openstack-network-exporter" containerID="cri-o://a2d8764d48f0504acb3368e626f02f443813a1caafd1fee9a37fbf6015805564" gracePeriod=300 Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.195426 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.204224 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.222445 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.247180 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.247265 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data podName:2559de32-42b0-4be1-b8ea-f42383d892a3 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:35.747247705 +0000 UTC m=+1526.537273285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data") pod "rabbitmq-cell1-server-0" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3") : configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.283269 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgx8r\" (UniqueName: \"kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r\") pod \"barbican-api-795dfbf48b-x8t2t\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.376021 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.376365 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data podName:03d954db-7dc6-4921-b260-1c189b9492c2 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:36.376347163 +0000 UTC m=+1527.166372743 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data") pod "rabbitmq-server-0" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2") : configmap "rabbitmq-config-data" not found Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.407393 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.444438 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.475089 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.475507 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="openstack-network-exporter" containerID="cri-o://50a79f09910e9fff54369d71bc6811a462f4e436d855283ce8d406341063c274" gracePeriod=300 Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.503537 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.576143 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.577826 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg7cl\" (UniqueName: \"kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl\") pod \"neutronec71-account-delete-8f8xz\" (UID: \"29980716-4f02-4c60-9116-e0c8f9615727\") " pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.679140 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg7cl\" (UniqueName: \"kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl\") pod \"neutronec71-account-delete-8f8xz\" (UID: \"29980716-4f02-4c60-9116-e0c8f9615727\") " pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.686913 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="ovsdbserver-sb" containerID="cri-o://c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" gracePeriod=300 Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.691401 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.692535 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.705767 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="ovsdbserver-nb" containerID="cri-o://77075241e0c723fffbcdd5be38b37be16523bbbe9459476084190610319b44be" gracePeriod=300 Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.745505 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:35 crc kubenswrapper[4989]: I1006 09:04:35.762005 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg7cl\" (UniqueName: \"kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl\") pod \"neutronec71-account-delete-8f8xz\" (UID: \"29980716-4f02-4c60-9116-e0c8f9615727\") " pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.797355 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:35 crc kubenswrapper[4989]: E1006 09:04:35.797548 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data podName:2559de32-42b0-4be1-b8ea-f42383d892a3 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:36.797508741 +0000 UTC m=+1527.587534311 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data") pod "rabbitmq-cell1-server-0" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3") : configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:35.883479 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:35.966073 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dp84\" (UniqueName: \"kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84\") pod \"cinderf526-account-delete-nt44j\" (UID: \"c6fa584c-5e07-489b-aac3-a91c35981b06\") " pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.051271 4989 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/swift-storage-0" secret="" err="secret \"swift-swift-dockercfg-qszr8\" not found" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.068126 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dp84\" (UniqueName: \"kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84\") pod \"cinderf526-account-delete-nt44j\" (UID: \"c6fa584c-5e07-489b-aac3-a91c35981b06\") " pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.145961 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dp84\" (UniqueName: \"kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84\") pod \"cinderf526-account-delete-nt44j\" (UID: \"c6fa584c-5e07-489b-aac3-a91c35981b06\") " pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.170684 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.170709 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.170721 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.170736 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.170781 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:36.670763048 +0000 UTC m=+1527.460788628 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.184068 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.185414 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_39fb4b45-4966-4e65-b368-322d8c65e0c8/ovsdbserver-nb/0.log" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.188400 4989 generic.go:334] "Generic (PLEG): container finished" podID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerID="77075241e0c723fffbcdd5be38b37be16523bbbe9459476084190610319b44be" exitCode=143 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196318 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196412 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4mdrv"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196468 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerDied","Data":"77075241e0c723fffbcdd5be38b37be16523bbbe9459476084190610319b44be"} Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196498 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196556 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4mdrv"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.196580 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6tthk"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.197735 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.198163 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" containerID="cri-o://b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.198314 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="openstack-network-exporter" containerID="cri-o://5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.198918 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ec3e388-5714-48ab-86a5-01962c90587f/ovsdbserver-sb/0.log" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.198992 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ec3e388-5714-48ab-86a5-01962c90587f" containerID="50a79f09910e9fff54369d71bc6811a462f4e436d855283ce8d406341063c274" exitCode=2 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.199048 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ec3e388-5714-48ab-86a5-01962c90587f" containerID="c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" exitCode=143 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.199116 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerDied","Data":"50a79f09910e9fff54369d71bc6811a462f4e436d855283ce8d406341063c274"} Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.199179 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerDied","Data":"c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7"} Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.224293 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.244545 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="dnsmasq-dns" containerID="cri-o://57bccaac485a945bfa34ccfa394125f43a85083ec91573de318dfac2c1dbab2e" gracePeriod=10 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.266698 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.317725 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6tthk"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.360762 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.361034 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-spqbp" podUID="2445aa06-1e2f-41fe-ae4d-a998f7874955" containerName="openstack-network-exporter" containerID="cri-o://d527d1e974bd70387c4d59349474b57da4a2251238da24621c65f48e2aba78db" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.373889 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9td6\" (UniqueName: \"kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6\") pod \"glanceec63-account-delete-5qcmq\" (UID: \"da55e23b-3ed8-41dc-875e-95cad75b9a19\") " pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:36 crc kubenswrapper[4989]: W1006 09:04:36.384594 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37d4951c_2719_4670_823e_f75163b3f472.slice/crio-ea2ceaf34d5fb5c8c9f8618b2a32ab47ebe35c68225b35d9a38c2cc03f49fd2d WatchSource:0}: Error finding container ea2ceaf34d5fb5c8c9f8618b2a32ab47ebe35c68225b35d9a38c2cc03f49fd2d: Status 404 returned error can't find the container with id ea2ceaf34d5fb5c8c9f8618b2a32ab47ebe35c68225b35d9a38c2cc03f49fd2d Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.414231 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-qkp2p"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.453833 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jl67c"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.477137 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9td6\" (UniqueName: \"kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6\") pod \"glanceec63-account-delete-5qcmq\" (UID: \"da55e23b-3ed8-41dc-875e-95cad75b9a19\") " pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.477732 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.479181 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data podName:03d954db-7dc6-4921-b260-1c189b9492c2 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:38.479029805 +0000 UTC m=+1529.269055385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data") pod "rabbitmq-server-0" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2") : configmap "rabbitmq-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.506833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9td6\" (UniqueName: \"kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6\") pod \"glanceec63-account-delete-5qcmq\" (UID: \"da55e23b-3ed8-41dc-875e-95cad75b9a19\") " pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.524725 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.564328 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-qkp2p"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.599670 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jl67c"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.616289 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-mwhkn"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.635861 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-v76fj"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.647715 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.649021 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.671822 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.672700 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-server" containerID="cri-o://15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673165 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="swift-recon-cron" containerID="cri-o://f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673232 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="rsync" containerID="cri-o://ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673287 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-expirer" containerID="cri-o://8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673339 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-updater" containerID="cri-o://3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673390 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-auditor" containerID="cri-o://6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673428 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-replicator" containerID="cri-o://662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673472 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-server" containerID="cri-o://0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673516 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-updater" containerID="cri-o://4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673580 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-auditor" containerID="cri-o://6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673633 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-replicator" containerID="cri-o://02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673705 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-server" containerID="cri-o://855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.673758 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-reaper" containerID="cri-o://e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.674210 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-auditor" containerID="cri-o://e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.674273 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-replicator" containerID="cri-o://9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.683585 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-xdjg5"] Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.686957 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.686984 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.686992 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.687002 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.687036 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:37.687023844 +0000 UTC m=+1528.477049424 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.726569 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-mwhkn"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.735331 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-xdjg5"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.746784 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-v76fj"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.753628 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.766463 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.777304 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-wtzcf"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.782194 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-wtzcf"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.790681 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxx2g\" (UniqueName: \"kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g\") pod \"novacell0ef11-account-delete-sjgsl\" (UID: \"b2a30a5a-7506-4880-9813-7aef02c9ca57\") " pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.797075 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.806715 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.823595 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.898700 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf7vs\" (UniqueName: \"kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs\") pod \"barbican8c94-account-delete-nh5mr\" (UID: \"8a895ac5-a669-43f0-85d5-071adb1d9762\") " pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.898880 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxx2g\" (UniqueName: \"kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g\") pod \"novacell0ef11-account-delete-sjgsl\" (UID: \"b2a30a5a-7506-4880-9813-7aef02c9ca57\") " pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.899558 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: E1006 09:04:36.899600 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data podName:2559de32-42b0-4be1-b8ea-f42383d892a3 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:38.899588325 +0000 UTC m=+1529.689613905 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data") pod "rabbitmq-cell1-server-0" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3") : configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.913350 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.924690 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5d88db6f8d-sz2w2" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-log" containerID="cri-o://35ffcc7392dc41bc132668ec57014e8ac2e8039a4359f8333605eabdde751c1e" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.925274 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5d88db6f8d-sz2w2" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-api" containerID="cri-o://4a55740396b636c8de2207c1b20cbc61bd80b8803c4aaa8ddc33405ea2cecaed" gracePeriod=30 Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.940972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxx2g\" (UniqueName: \"kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g\") pod \"novacell0ef11-account-delete-sjgsl\" (UID: \"b2a30a5a-7506-4880-9813-7aef02c9ca57\") " pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.956947 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:36 crc kubenswrapper[4989]: I1006 09:04:36.971532 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.002267 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf7vs\" (UniqueName: \"kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs\") pod \"barbican8c94-account-delete-nh5mr\" (UID: \"8a895ac5-a669-43f0-85d5-071adb1d9762\") " pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.067336 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.106168 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.107875 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf7vs\" (UniqueName: \"kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs\") pod \"barbican8c94-account-delete-nh5mr\" (UID: \"8a895ac5-a669-43f0-85d5-071adb1d9762\") " pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.108635 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.126061 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.129051 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.129562 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-557b4f9dd9-blscg" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-api" containerID="cri-o://4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.133704 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-557b4f9dd9-blscg" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-httpd" containerID="cri-o://85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.146533 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.146821 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-httpd" containerID="cri-o://4368459492d84638518335c2551dd210d155262a4a5fb3137e085d2b4dab20cc" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.147414 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-server" containerID="cri-o://24bd4b9578823a45ed78764215744ad5b434e9e908686d6d508220f295123fc5" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.157723 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ca6e-account-create-c8xvj"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.174738 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ca6e-account-create-c8xvj"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.186221 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zs457"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.186642 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zs457"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.193866 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ec71-account-create-ms5qt"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.200713 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ec71-account-create-ms5qt"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.205893 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nf448"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.213464 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nf448"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.219082 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.232235 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.239386 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.247102 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerStarted","Data":"ea2ceaf34d5fb5c8c9f8618b2a32ab47ebe35c68225b35d9a38c2cc03f49fd2d"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.251138 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.251424 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-log" containerID="cri-o://3814cda71279e66c461650e21aeab490084205cdc04eb79209f77094180bab6e" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.251478 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-httpd" containerID="cri-o://a7ea9a251ec17f10791afe94beed6956bf60120b934e4c1b26a49ceedb4b0502" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.254937 4989 generic.go:334] "Generic (PLEG): container finished" podID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerID="5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d" exitCode=2 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.255035 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerDied","Data":"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.270388 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-spqbp_2445aa06-1e2f-41fe-ae4d-a998f7874955/openstack-network-exporter/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.270436 4989 generic.go:334] "Generic (PLEG): container finished" podID="2445aa06-1e2f-41fe-ae4d-a998f7874955" containerID="d527d1e974bd70387c4d59349474b57da4a2251238da24621c65f48e2aba78db" exitCode=2 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.270533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-spqbp" event={"ID":"2445aa06-1e2f-41fe-ae4d-a998f7874955","Type":"ContainerDied","Data":"d527d1e974bd70387c4d59349474b57da4a2251238da24621c65f48e2aba78db"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.272627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerStarted","Data":"7ec3f01a0b55542dcc99e22e3799f8d5c190928fd4f5bdccdf80b385eaafeaad"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.287871 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.288158 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-log" containerID="cri-o://9e2d55d5310b2a032c3053d2f611de15d5d3b47bc88f9275f2f82580ad17ce65" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.288326 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-api" containerID="cri-o://de97aa68267d73cae2bc02e80acd84df5900e6e32436971bd32cc722e73765e7" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.293504 4989 generic.go:334] "Generic (PLEG): container finished" podID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerID="e2d916bd8ce663a1374a2102f14126cf7f39b34e74293d78b95861c7f07ea483" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.293588 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerDied","Data":"e2d916bd8ce663a1374a2102f14126cf7f39b34e74293d78b95861c7f07ea483"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.307581 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_39fb4b45-4966-4e65-b368-322d8c65e0c8/ovsdbserver-nb/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.307624 4989 generic.go:334] "Generic (PLEG): container finished" podID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerID="a2d8764d48f0504acb3368e626f02f443813a1caafd1fee9a37fbf6015805564" exitCode=2 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.307682 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerDied","Data":"a2d8764d48f0504acb3368e626f02f443813a1caafd1fee9a37fbf6015805564"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.328810 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-2n8b2"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.339323 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" containerID="cri-o://9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" gracePeriod=29 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.350544 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-2n8b2"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.360925 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.361143 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" containerID="cri-o://648b0a633bb695eac1b4b106abd3e4423a4c45eceef2c1366e8d7b5a4a2862d4" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.361562 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" containerID="cri-o://41c3a3ea88eda96a1475512601b4880062d4b09816817a60f531e6bc6a058912" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.404110 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f526-account-create-t8hsg"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405494 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405519 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405529 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405539 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405547 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405554 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405562 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405569 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405577 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405584 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405591 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405686 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405716 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405742 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405752 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405764 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405775 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405785 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405796 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405807 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.405818 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.412564 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f526-account-create-t8hsg"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.416121 4989 generic.go:334] "Generic (PLEG): container finished" podID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerID="57bccaac485a945bfa34ccfa394125f43a85083ec91573de318dfac2c1dbab2e" exitCode=0 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.416264 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" event={"ID":"80aa68cc-125e-4753-8c49-17f39da3e3f5","Type":"ContainerDied","Data":"57bccaac485a945bfa34ccfa394125f43a85083ec91573de318dfac2c1dbab2e"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.416299 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" event={"ID":"80aa68cc-125e-4753-8c49-17f39da3e3f5","Type":"ContainerDied","Data":"a67edcbcec88bf76f64b2e568efe9e42177d529c33d9a9979cd05231aed7ef2d"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.416314 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a67edcbcec88bf76f64b2e568efe9e42177d529c33d9a9979cd05231aed7ef2d" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.420874 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.441937 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.441982 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="rabbitmq" containerID="cri-o://97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032" gracePeriod=604800 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.442271 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-log" containerID="cri-o://1503150658fd3c0f67b76860cf73488324464b757cbc4426f8c36b6c4913eb11" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.442416 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-httpd" containerID="cri-o://9e3f96ebea5bce42b75defb3a323c9cc21c5adae1bd4c054f3fd804d7a885721" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.463989 4989 generic.go:334] "Generic (PLEG): container finished" podID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" containerID="ca2377b7cd7f73d36c389bf39d5a5b793d4492f8b024e80813557074b081fb4c" exitCode=137 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.484156 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.488632 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerStarted","Data":"22318f22bce93e507587fad2cc347182d920498706fced4191b2e168c632aeca"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.490761 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ec3e388-5714-48ab-86a5-01962c90587f/ovsdbserver-sb/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.490833 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1ec3e388-5714-48ab-86a5-01962c90587f","Type":"ContainerDied","Data":"cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.490861 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb7ebcfc391e5da628f7c8f6a04c500c000459427edfe9dd2e7ce8d3c0255eec" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.506156 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5fmhh"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.508913 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="rabbitmq" containerID="cri-o://e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca" gracePeriod=604800 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.515874 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerID="35ffcc7392dc41bc132668ec57014e8ac2e8039a4359f8333605eabdde751c1e" exitCode=143 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.515915 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerDied","Data":"35ffcc7392dc41bc132668ec57014e8ac2e8039a4359f8333605eabdde751c1e"} Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.530161 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5fmhh"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.530647 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7 is running failed: container process not found" containerID="c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.531282 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7 is running failed: container process not found" containerID="c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.533075 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7 is running failed: container process not found" containerID="c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.533107 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="ovsdbserver-sb" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.542689 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4c6e-account-create-zhzb5"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.569946 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4c6e-account-create-zhzb5"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.589231 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.589540 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://fa81e6c26183ea833e5e2d25e80e937b2130d7c82f3b53456b03c4b55db2e255" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.599210 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-jfdd5"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.639364 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-jfdd5"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.642905 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-ded3-account-create-2p5m9"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.650810 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.651173 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker-log" containerID="cri-o://206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.652778 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker" containerID="cri-o://79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.657136 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6494d"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.668814 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6494d"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.688577 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-ded3-account-create-2p5m9"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.718236 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.732707 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.741944 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.741976 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.741985 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.741995 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.742036 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:39.742023584 +0000 UTC m=+1530.532049164 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.753031 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.753271 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener-log" containerID="cri-o://fac01ec109fb091c49f98d9c013903a2e4a8fcce4fac1a9f3b69848da9287623" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.753742 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener" containerID="cri-o://ca7e527853be7d729e32e29acc9093bf1307493ccd1757e678492f80c2ca1c5e" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.765728 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec63-account-create-2r8tq"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.781828 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec63-account-create-2r8tq"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.790460 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-9mh59"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.804159 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.810935 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-9mh59"] Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.851430 4989 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 06 09:04:37 crc kubenswrapper[4989]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 06 09:04:37 crc kubenswrapper[4989]: + source /usr/local/bin/container-scripts/functions Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNBridge=br-int Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNRemote=tcp:localhost:6642 Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNEncapType=geneve Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNAvailabilityZones= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ EnableChassisAsGateway=true Oct 06 09:04:37 crc kubenswrapper[4989]: ++ PhysicalNetworks= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNHostName= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 06 09:04:37 crc kubenswrapper[4989]: ++ ovs_dir=/var/lib/openvswitch Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 06 09:04:37 crc kubenswrapper[4989]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + cleanup_ovsdb_server_semaphore Oct 06 09:04:37 crc kubenswrapper[4989]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 06 09:04:37 crc kubenswrapper[4989]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-tmq59" message=< Oct 06 09:04:37 crc kubenswrapper[4989]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 06 09:04:37 crc kubenswrapper[4989]: + source /usr/local/bin/container-scripts/functions Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNBridge=br-int Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNRemote=tcp:localhost:6642 Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNEncapType=geneve Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNAvailabilityZones= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ EnableChassisAsGateway=true Oct 06 09:04:37 crc kubenswrapper[4989]: ++ PhysicalNetworks= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNHostName= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 06 09:04:37 crc kubenswrapper[4989]: ++ ovs_dir=/var/lib/openvswitch Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 06 09:04:37 crc kubenswrapper[4989]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + cleanup_ovsdb_server_semaphore Oct 06 09:04:37 crc kubenswrapper[4989]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 06 09:04:37 crc kubenswrapper[4989]: > Oct 06 09:04:37 crc kubenswrapper[4989]: E1006 09:04:37.851490 4989 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 06 09:04:37 crc kubenswrapper[4989]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 06 09:04:37 crc kubenswrapper[4989]: + source /usr/local/bin/container-scripts/functions Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNBridge=br-int Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNRemote=tcp:localhost:6642 Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNEncapType=geneve Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNAvailabilityZones= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ EnableChassisAsGateway=true Oct 06 09:04:37 crc kubenswrapper[4989]: ++ PhysicalNetworks= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ OVNHostName= Oct 06 09:04:37 crc kubenswrapper[4989]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 06 09:04:37 crc kubenswrapper[4989]: ++ ovs_dir=/var/lib/openvswitch Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 06 09:04:37 crc kubenswrapper[4989]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 06 09:04:37 crc kubenswrapper[4989]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + sleep 0.5 Oct 06 09:04:37 crc kubenswrapper[4989]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 06 09:04:37 crc kubenswrapper[4989]: + cleanup_ovsdb_server_semaphore Oct 06 09:04:37 crc kubenswrapper[4989]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 06 09:04:37 crc kubenswrapper[4989]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 06 09:04:37 crc kubenswrapper[4989]: > pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" containerID="cri-o://8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.851532 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" containerID="cri-o://8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" gracePeriod=29 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.851963 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1ec3e388-5714-48ab-86a5-01962c90587f/ovsdbserver-sb/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.852044 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.862223 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-spqbp_2445aa06-1e2f-41fe-ae4d-a998f7874955/openstack-network-exporter/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.871635 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.867164 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.867071 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.873526 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.879622 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_39fb4b45-4966-4e65-b368-322d8c65e0c8/ovsdbserver-nb/0.log" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.879729 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.931769 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="galera" containerID="cri-o://c0e57d5d18f141af4479840759a41e612d6a1d950914d8c0d137ed2381bc7ce1" gracePeriod=30 Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953583 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953644 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953704 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953731 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953756 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953828 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953859 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953882 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953901 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953954 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.953993 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954039 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954056 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954082 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954112 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle\") pod \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954149 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954178 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcr4q\" (UniqueName: \"kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954200 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954226 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66gtt\" (UniqueName: \"kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954254 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sddhj\" (UniqueName: \"kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954284 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir\") pod \"2445aa06-1e2f-41fe-ae4d-a998f7874955\" (UID: \"2445aa06-1e2f-41fe-ae4d-a998f7874955\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954311 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxn6d\" (UniqueName: \"kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d\") pod \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954335 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954361 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config\") pod \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954392 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954417 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret\") pod \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\" (UID: \"629e7fe3-c002-4313-8c5b-bdf82e4adcd5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954441 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5kt4\" (UniqueName: \"kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954470 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954486 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1ec3e388-5714-48ab-86a5-01962c90587f\" (UID: \"1ec3e388-5714-48ab-86a5-01962c90587f\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954517 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0\") pod \"80aa68cc-125e-4753-8c49-17f39da3e3f5\" (UID: \"80aa68cc-125e-4753-8c49-17f39da3e3f5\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.954592 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle\") pod \"39fb4b45-4966-4e65-b368-322d8c65e0c8\" (UID: \"39fb4b45-4966-4e65-b368-322d8c65e0c8\") " Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.958168 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config" (OuterVolumeSpecName: "config") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.971122 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config" (OuterVolumeSpecName: "config") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.975058 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.976378 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts" (OuterVolumeSpecName: "scripts") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.979640 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config" (OuterVolumeSpecName: "config") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.979719 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.980279 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.980866 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.986216 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts" (OuterVolumeSpecName: "scripts") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:37 crc kubenswrapper[4989]: I1006 09:04:37.998119 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt" (OuterVolumeSpecName: "kube-api-access-66gtt") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "kube-api-access-66gtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.000246 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02a2fb2b-b301-444d-9e4b-f8105f3f229e" path="/var/lib/kubelet/pods/02a2fb2b-b301-444d-9e4b-f8105f3f229e/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.001138 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf" path="/var/lib/kubelet/pods/0428e6e7-069a-4ae7-9ec3-bfc37fed9ebf/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.001780 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d66a34-26bc-4c9a-b0e6-d4bc809507f9" path="/var/lib/kubelet/pods/04d66a34-26bc-4c9a-b0e6-d4bc809507f9/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.002588 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bbd7317-2922-431a-bbf2-a22515821224" path="/var/lib/kubelet/pods/1bbd7317-2922-431a-bbf2-a22515821224/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.003738 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20284027-d662-48a8-b61c-3b089247c1ea" path="/var/lib/kubelet/pods/20284027-d662-48a8-b61c-3b089247c1ea/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.004290 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2137ce1b-b046-4b3b-ab0a-3566df04dd46" path="/var/lib/kubelet/pods/2137ce1b-b046-4b3b-ab0a-3566df04dd46/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.005008 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4" (OuterVolumeSpecName: "kube-api-access-g5kt4") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "kube-api-access-g5kt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.005031 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="278f3372-6bb1-4ae2-871a-80a084a21c8d" path="/var/lib/kubelet/pods/278f3372-6bb1-4ae2-871a-80a084a21c8d/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.007852 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b32cd31-074d-4d27-aae2-dbcc01c6810b" path="/var/lib/kubelet/pods/3b32cd31-074d-4d27-aae2-dbcc01c6810b/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.008526 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d14187c-d5ec-4f67-b648-8d3d99e833c1" path="/var/lib/kubelet/pods/4d14187c-d5ec-4f67-b648-8d3d99e833c1/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.009115 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604e87ea-9072-4967-b6b8-c717378a4db7" path="/var/lib/kubelet/pods/604e87ea-9072-4967-b6b8-c717378a4db7/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.014710 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9edea7-f372-42ff-bd54-bb1a919f79ac" path="/var/lib/kubelet/pods/6d9edea7-f372-42ff-bd54-bb1a919f79ac/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.015670 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73589ab3-b75d-4d92-b312-2464ba72317a" path="/var/lib/kubelet/pods/73589ab3-b75d-4d92-b312-2464ba72317a/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.016328 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86625194-ed32-446c-869a-d1a8d30a5c93" path="/var/lib/kubelet/pods/86625194-ed32-446c-869a-d1a8d30a5c93/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.019002 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f6adc1f-6029-46fe-939c-8633e05cb14b" path="/var/lib/kubelet/pods/8f6adc1f-6029-46fe-939c-8633e05cb14b/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.021286 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a144d611-7769-46ce-8707-3814fb0a7aa7" path="/var/lib/kubelet/pods/a144d611-7769-46ce-8707-3814fb0a7aa7/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.022428 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q" (OuterVolumeSpecName: "kube-api-access-jcr4q") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "kube-api-access-jcr4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.027810 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4fe38d-017e-40fa-8d24-3da65f03839d" path="/var/lib/kubelet/pods/ab4fe38d-017e-40fa-8d24-3da65f03839d/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.028745 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b416f76a-8dcd-411a-a0a0-31f7becea55e" path="/var/lib/kubelet/pods/b416f76a-8dcd-411a-a0a0-31f7becea55e/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.029426 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b65de4a0-67b5-446e-896a-6934b9d11e9b" path="/var/lib/kubelet/pods/b65de4a0-67b5-446e-896a-6934b9d11e9b/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.030298 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.032256 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d" (OuterVolumeSpecName: "kube-api-access-fxn6d") pod "629e7fe3-c002-4313-8c5b-bdf82e4adcd5" (UID: "629e7fe3-c002-4313-8c5b-bdf82e4adcd5"). InnerVolumeSpecName "kube-api-access-fxn6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.032392 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.036856 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj" (OuterVolumeSpecName: "kube-api-access-sddhj") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "kube-api-access-sddhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058226 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058246 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2445aa06-1e2f-41fe-ae4d-a998f7874955-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058255 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058264 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ec3e388-5714-48ab-86a5-01962c90587f-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058272 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcr4q\" (UniqueName: \"kubernetes.io/projected/2445aa06-1e2f-41fe-ae4d-a998f7874955-kube-api-access-jcr4q\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058282 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66gtt\" (UniqueName: \"kubernetes.io/projected/80aa68cc-125e-4753-8c49-17f39da3e3f5-kube-api-access-66gtt\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058290 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sddhj\" (UniqueName: \"kubernetes.io/projected/1ec3e388-5714-48ab-86a5-01962c90587f-kube-api-access-sddhj\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058298 4989 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058306 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxn6d\" (UniqueName: \"kubernetes.io/projected/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-kube-api-access-fxn6d\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058314 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058321 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5kt4\" (UniqueName: \"kubernetes.io/projected/39fb4b45-4966-4e65-b368-322d8c65e0c8-kube-api-access-g5kt4\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058329 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39fb4b45-4966-4e65-b368-322d8c65e0c8-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058352 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058361 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2445aa06-1e2f-41fe-ae4d-a998f7874955-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058369 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.058381 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.069838 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d81dcb4a-9801-4a2d-a106-0b7916a6b569" path="/var/lib/kubelet/pods/d81dcb4a-9801-4a2d-a106-0b7916a6b569/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.071013 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c" path="/var/lib/kubelet/pods/e5c7c63e-55e8-47f0-bf26-f5b83ea54b4c/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.071787 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eef21e7e-a1fe-4525-8bbe-5a7adac861cb" path="/var/lib/kubelet/pods/eef21e7e-a1fe-4525-8bbe-5a7adac861cb/volumes" Oct 06 09:04:38 crc kubenswrapper[4989]: W1006 09:04:38.192354 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6fa584c_5e07_489b_aac3_a91c35981b06.slice/crio-0861bb01ae8435110462fe291366e108fc5b9d6d89ecf0d5b87fdf43199716b0 WatchSource:0}: Error finding container 0861bb01ae8435110462fe291366e108fc5b9d6d89ecf0d5b87fdf43199716b0: Status 404 returned error can't find the container with id 0861bb01ae8435110462fe291366e108fc5b9d6d89ecf0d5b87fdf43199716b0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.343026 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.373626 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: W1006 09:04:38.509222 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda55e23b_3ed8_41dc_875e_95cad75b9a19.slice/crio-7ba3bd59092f163e1a214bdb9280ce107e2e2ee6424baf9c75cd83b5c7fcd188 WatchSource:0}: Error finding container 7ba3bd59092f163e1a214bdb9280ce107e2e2ee6424baf9c75cd83b5c7fcd188: Status 404 returned error can't find the container with id 7ba3bd59092f163e1a214bdb9280ce107e2e2ee6424baf9c75cd83b5c7fcd188 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.572937 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.577192 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: E1006 09:04:38.577251 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 06 09:04:38 crc kubenswrapper[4989]: E1006 09:04:38.577286 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data podName:03d954db-7dc6-4921-b260-1c189b9492c2 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:42.577275626 +0000 UTC m=+1533.367301206 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data") pod "rabbitmq-server-0" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2") : configmap "rabbitmq-config-data" not found Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.578094 4989 generic.go:334] "Generic (PLEG): container finished" podID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerID="f4b25eb0db619fd2778bd141569f6d60f1ea4b0b8e5c38cb07786cb6a462723c" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.584529 4989 generic.go:334] "Generic (PLEG): container finished" podID="29637231-ec41-4e57-af1f-18046cc06cf6" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.586729 4989 generic.go:334] "Generic (PLEG): container finished" podID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerID="206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.603673 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.603701 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.603709 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: W1006 09:04:38.618894 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a895ac5_a669_43f0_85d5_071adb1d9762.slice/crio-780521d1c60ffd8a7a82e20039a146e6d64b70dcc354c248fd0125e83ba58888 WatchSource:0}: Error finding container 780521d1c60ffd8a7a82e20039a146e6d64b70dcc354c248fd0125e83ba58888: Status 404 returned error can't find the container with id 780521d1c60ffd8a7a82e20039a146e6d64b70dcc354c248fd0125e83ba58888 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.625059 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_39fb4b45-4966-4e65-b368-322d8c65e0c8/ovsdbserver-nb/0.log" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.625194 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.642699 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.669014 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerID="fac01ec109fb091c49f98d9c013903a2e4a8fcce4fac1a9f3b69848da9287623" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.679291 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.736255 4989 generic.go:334] "Generic (PLEG): container finished" podID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerID="3814cda71279e66c461650e21aeab490084205cdc04eb79209f77094180bab6e" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.738438 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68c46dd79-4twxt" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker-log" containerID="cri-o://7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed" gracePeriod=30 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.740792 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-68c46dd79-4twxt" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker" containerID="cri-o://66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2" gracePeriod=30 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.750312 4989 generic.go:334] "Generic (PLEG): container finished" podID="74245807-f752-4209-a489-d3e746b5e1fb" containerID="1503150658fd3c0f67b76860cf73488324464b757cbc4426f8c36b6c4913eb11" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.754125 4989 generic.go:334] "Generic (PLEG): container finished" podID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerID="85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.756532 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.758494 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "629e7fe3-c002-4313-8c5b-bdf82e4adcd5" (UID: "629e7fe3-c002-4313-8c5b-bdf82e4adcd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.770112 4989 generic.go:334] "Generic (PLEG): container finished" podID="c29eee10-1638-43c9-806c-9c956d10f31a" containerID="9e2d55d5310b2a032c3053d2f611de15d5d3b47bc88f9275f2f82580ad17ce65" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.771328 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68c46dd79-4twxt" podStartSLOduration=4.771311024 podStartE2EDuration="4.771311024s" podCreationTimestamp="2025-10-06 09:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:04:38.768098742 +0000 UTC m=+1529.558124322" watchObservedRunningTime="2025-10-06 09:04:38.771311024 +0000 UTC m=+1529.561336604" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.776708 4989 generic.go:334] "Generic (PLEG): container finished" podID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerID="fa81e6c26183ea833e5e2d25e80e937b2130d7c82f3b53456b03c4b55db2e255" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.781084 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.793345 4989 generic.go:334] "Generic (PLEG): container finished" podID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerID="648b0a633bb695eac1b4b106abd3e4423a4c45eceef2c1366e8d7b5a4a2862d4" exitCode=143 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.812253 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.819926 4989 generic.go:334] "Generic (PLEG): container finished" podID="63329d01-3db3-448a-b22c-53a0d58f8497" containerID="24bd4b9578823a45ed78764215744ad5b434e9e908686d6d508220f295123fc5" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.819996 4989 generic.go:334] "Generic (PLEG): container finished" podID="63329d01-3db3-448a-b22c-53a0d58f8497" containerID="4368459492d84638518335c2551dd210d155262a4a5fb3137e085d2b4dab20cc" exitCode=0 Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.854814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.867376 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-spqbp_2445aa06-1e2f-41fe-ae4d-a998f7874955/openstack-network-exporter/0.log" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.867770 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.867801 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-spqbp" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.867849 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-bmrx4" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.882553 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.882692 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.962914 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.973675 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "629e7fe3-c002-4313-8c5b-bdf82e4adcd5" (UID: "629e7fe3-c002-4313-8c5b-bdf82e4adcd5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.984447 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: I1006 09:04:38.984475 4989 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:38 crc kubenswrapper[4989]: E1006 09:04:38.984529 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:38 crc kubenswrapper[4989]: E1006 09:04:38.984572 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data podName:2559de32-42b0-4be1-b8ea-f42383d892a3 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:42.984557855 +0000 UTC m=+1533.774583435 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data") pod "rabbitmq-cell1-server-0" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3") : configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.004966 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.032189 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config" (OuterVolumeSpecName: "config") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.085933 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.125592 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "629e7fe3-c002-4313-8c5b-bdf82e4adcd5" (UID: "629e7fe3-c002-4313-8c5b-bdf82e4adcd5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.136947 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.194497 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/629e7fe3-c002-4313-8c5b-bdf82e4adcd5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.194536 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.240329 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.259057 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "80aa68cc-125e-4753-8c49-17f39da3e3f5" (UID: "80aa68cc-125e-4753-8c49-17f39da3e3f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.294241 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.299827 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.299860 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.299871 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80aa68cc-125e-4753-8c49-17f39da3e3f5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.325836 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.329939 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2445aa06-1e2f-41fe-ae4d-a998f7874955" (UID: "2445aa06-1e2f-41fe-ae4d-a998f7874955"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.385745 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "1ec3e388-5714-48ab-86a5-01962c90587f" (UID: "1ec3e388-5714-48ab-86a5-01962c90587f"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.401560 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2445aa06-1e2f-41fe-ae4d-a998f7874955-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.401868 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ec3e388-5714-48ab-86a5-01962c90587f-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.401877 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.404166 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "39fb4b45-4966-4e65-b368-322d8c65e0c8" (UID: "39fb4b45-4966-4e65-b368-322d8c65e0c8"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.428812 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": read tcp 10.217.0.2:51528->10.217.0.161:8776: read: connection reset by peer" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.504273 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/39fb4b45-4966-4e65-b368-322d8c65e0c8-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.699397 4989 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.764s" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerDied","Data":"f4b25eb0db619fd2778bd141569f6d60f1ea4b0b8e5c38cb07786cb6a462723c"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699554 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerDied","Data":"8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699574 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ef11-account-create-lp9dg"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699594 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ef11-account-create-lp9dg"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699610 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lmkzl"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699621 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerDied","Data":"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699649 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699678 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8c94-account-create-zhkkc"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699691 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699701 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699718 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lmkzl"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699730 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8c94-account-create-zhkkc"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699745 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf526-account-delete-nt44j" event={"ID":"c6fa584c-5e07-489b-aac3-a91c35981b06","Type":"ContainerStarted","Data":"0861bb01ae8435110462fe291366e108fc5b9d6d89ecf0d5b87fdf43199716b0"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699757 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699772 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"39fb4b45-4966-4e65-b368-322d8c65e0c8","Type":"ContainerDied","Data":"bfe39e3bdd8a5199ebb85e3e073a40a85088fc0bd159a7d2427941258c551a11"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699787 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699803 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699821 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef11-account-delete-sjgsl" event={"ID":"b2a30a5a-7506-4880-9813-7aef02c9ca57","Type":"ContainerStarted","Data":"3708f56e2bd61189da05b826d96da515af7aba3ef8b112e5c9c170019698e9d6"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699835 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerDied","Data":"fac01ec109fb091c49f98d9c013903a2e4a8fcce4fac1a9f3b69848da9287623"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699850 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceec63-account-delete-5qcmq" event={"ID":"da55e23b-3ed8-41dc-875e-95cad75b9a19","Type":"ContainerStarted","Data":"7ba3bd59092f163e1a214bdb9280ce107e2e2ee6424baf9c75cd83b5c7fcd188"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.699863 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700182 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700194 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700212 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="dnsmasq-dns" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700220 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="dnsmasq-dns" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700244 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2445aa06-1e2f-41fe-ae4d-a998f7874955" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700252 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2445aa06-1e2f-41fe-ae4d-a998f7874955" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700269 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="ovsdbserver-sb" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700277 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="ovsdbserver-sb" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700291 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700297 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700307 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="ovsdbserver-nb" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700313 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="ovsdbserver-nb" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.700323 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="init" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700328 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="init" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700508 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700524 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="ovsdbserver-sb" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700533 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700546 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" containerName="dnsmasq-dns" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700554 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" containerName="ovsdbserver-nb" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.700563 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2445aa06-1e2f-41fe-ae4d-a998f7874955" containerName="openstack-network-exporter" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702073 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702097 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-r7w8z"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702109 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerStarted","Data":"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702129 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerDied","Data":"3814cda71279e66c461650e21aeab490084205cdc04eb79209f77094180bab6e"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702145 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerStarted","Data":"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702157 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerStarted","Data":"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702166 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerDied","Data":"1503150658fd3c0f67b76860cf73488324464b757cbc4426f8c36b6c4913eb11"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702182 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-r7w8z"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702196 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerDied","Data":"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702210 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702228 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronec71-account-delete-8f8xz" event={"ID":"29980716-4f02-4c60-9116-e0c8f9615727","Type":"ContainerStarted","Data":"17987688491a57d4a92caa2a16b302757802609966c2e371a7024efac9c878a1"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702239 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerDied","Data":"9e2d55d5310b2a032c3053d2f611de15d5d3b47bc88f9275f2f82580ad17ce65"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702252 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5lp9k"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702265 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5lp9k"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702276 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702288 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702459 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" containerName="nova-cell0-conductor-conductor" containerID="cri-o://275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702488 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702520 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50fc7e8a-a4c4-43ca-b672-7d8f63e25738","Type":"ContainerDied","Data":"fa81e6c26183ea833e5e2d25e80e937b2130d7c82f3b53456b03c4b55db2e255"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702542 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702553 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerDied","Data":"648b0a633bb695eac1b4b106abd3e4423a4c45eceef2c1366e8d7b5a4a2862d4"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702566 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerDied","Data":"24bd4b9578823a45ed78764215744ad5b434e9e908686d6d508220f295123fc5"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702577 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702588 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702605 4989 scope.go:117] "RemoveContainer" containerID="a2d8764d48f0504acb3368e626f02f443813a1caafd1fee9a37fbf6015805564" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702625 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerDied","Data":"4368459492d84638518335c2551dd210d155262a4a5fb3137e085d2b4dab20cc"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerStarted","Data":"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.702761 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-spqbp" event={"ID":"2445aa06-1e2f-41fe-ae4d-a998f7874955","Type":"ContainerDied","Data":"cbb6d3fbedb7c7e24b23449713d9d60e1f50002adb80d74c5257ca74cf66e50d"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.703078 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57c669566-xd94v" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api-log" containerID="cri-o://c0e526eeb3fc63bca07ba53b274454c122438ef8d646f58978c54082302aad38" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.703452 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.703557 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerName="nova-scheduler-scheduler" containerID="cri-o://051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.703619 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57c669566-xd94v" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api" containerID="cri-o://aa79f2c4a47bfd9cd6e90a260e7452c6eaabae169b34d4f1882b65c8d80ca5ac" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.758165 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.760235 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.761397 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.761425 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.806911 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.807211 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.807618 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.808632 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.808721 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.809545 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.811383 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.811456 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.812527 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.812783 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8vrz\" (UniqueName: \"kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.812805 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.814093 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.814116 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.814123 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.814133 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:39 crc kubenswrapper[4989]: E1006 09:04:39.814172 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:43.814158073 +0000 UTC m=+1534.604183643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.866691 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.878971 4989 scope.go:117] "RemoveContainer" containerID="77075241e0c723fffbcdd5be38b37be16523bbbe9459476084190610319b44be" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.884723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerStarted","Data":"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.887670 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener-log" containerID="cri-o://1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.888033 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener" containerID="cri-o://87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3" gracePeriod=30 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.900738 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e68a18fb-33b8-4ddf-95a7-46563d1ed77e","Type":"ContainerDied","Data":"c7122261bc6367b1cee3cee833a5c28aad55d9f06208cc3a691b7d0742e6e521"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.901205 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7122261bc6367b1cee3cee833a5c28aad55d9f06208cc3a691b7d0742e6e521" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.903078 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"50fc7e8a-a4c4-43ca-b672-7d8f63e25738","Type":"ContainerDied","Data":"056cd8afe444128c40113c0f59d0428a85aa64ee07b6b59a9d92282bfabe65cb"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.903328 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="056cd8afe444128c40113c0f59d0428a85aa64ee07b6b59a9d92282bfabe65cb" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.913388 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" podStartSLOduration=5.91336871 podStartE2EDuration="5.91336871s" podCreationTimestamp="2025-10-06 09:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:04:39.908343965 +0000 UTC m=+1530.698369535" watchObservedRunningTime="2025-10-06 09:04:39.91336871 +0000 UTC m=+1530.703394290" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.917547 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.917584 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.917690 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.917736 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919136 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919208 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919236 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j95dt\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919280 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd\") pod \"63329d01-3db3-448a-b22c-53a0d58f8497\" (UID: \"63329d01-3db3-448a-b22c-53a0d58f8497\") " Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919561 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919707 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8vrz\" (UniqueName: \"kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.919732 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.920496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.921249 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.921211 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.921810 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8c94-account-delete-nh5mr" event={"ID":"8a895ac5-a669-43f0-85d5-071adb1d9762","Type":"ContainerStarted","Data":"780521d1c60ffd8a7a82e20039a146e6d64b70dcc354c248fd0125e83ba58888"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.928175 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.933140 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.933557 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" event={"ID":"63329d01-3db3-448a-b22c-53a0d58f8497","Type":"ContainerDied","Data":"8d21c1f0abb03bb68ac6359477e03082b1bf2546df6cdaadd81935a06b5a5193"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.938692 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt" (OuterVolumeSpecName: "kube-api-access-j95dt") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "kube-api-access-j95dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.942538 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8vrz\" (UniqueName: \"kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz\") pod \"redhat-operators-7fnl7\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.945826 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.948886 4989 generic.go:334] "Generic (PLEG): container finished" podID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerID="c0e57d5d18f141af4479840759a41e612d6a1d950914d8c0d137ed2381bc7ce1" exitCode=0 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.954804 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.968875 4989 generic.go:334] "Generic (PLEG): container finished" podID="bc313ada-f4bc-4711-838a-b809c4e29273" containerID="7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed" exitCode=143 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.971610 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093ba9d5-8d9d-4e0d-837b-21575e2ad9dd" path="/var/lib/kubelet/pods/093ba9d5-8d9d-4e0d-837b-21575e2ad9dd/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.972237 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da63f86-cdd5-4199-88a1-b617db5c046b" path="/var/lib/kubelet/pods/1da63f86-cdd5-4199-88a1-b617db5c046b/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.972951 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="227e2f8d-debb-4226-a415-73b3e9bedcd4" path="/var/lib/kubelet/pods/227e2f8d-debb-4226-a415-73b3e9bedcd4/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.973739 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629e7fe3-c002-4313-8c5b-bdf82e4adcd5" path="/var/lib/kubelet/pods/629e7fe3-c002-4313-8c5b-bdf82e4adcd5/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.973943 4989 generic.go:334] "Generic (PLEG): container finished" podID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerID="0bf97fec682ff1dbd5c988be1156a9e788940eb34663d369552ca96c1e405f14" exitCode=0 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.974780 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aec9ce1-0a3f-4788-a367-4fa2f1554d44" path="/var/lib/kubelet/pods/8aec9ce1-0a3f-4788-a367-4fa2f1554d44/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.976102 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e" path="/var/lib/kubelet/pods/9a87fd5d-ae89-4d9f-aa82-1ef0380d9f0e/volumes" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.979415 4989 generic.go:334] "Generic (PLEG): container finished" podID="4efffd9e-87ec-4569-be26-03091469d765" containerID="c0e526eeb3fc63bca07ba53b274454c122438ef8d646f58978c54082302aad38" exitCode=143 Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.993497 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997857 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerDied","Data":"c0e57d5d18f141af4479840759a41e612d6a1d950914d8c0d137ed2381bc7ce1"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997898 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8","Type":"ContainerDied","Data":"07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997910 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerDied","Data":"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997932 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerDied","Data":"0bf97fec682ff1dbd5c988be1156a9e788940eb34663d369552ca96c1e405f14"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997944 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2e1dc821-e800-4146-b43c-55f73af4daf0","Type":"ContainerDied","Data":"82dc5eabc22659eeb656fbffd76032ecf98f8043f68f61542947074b4d3da8f9"} Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997953 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82dc5eabc22659eeb656fbffd76032ecf98f8043f68f61542947074b4d3da8f9" Oct 06 09:04:39 crc kubenswrapper[4989]: I1006 09:04:39.997961 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerDied","Data":"c0e526eeb3fc63bca07ba53b274454c122438ef8d646f58978c54082302aad38"} Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.029476 4989 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.033417 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j95dt\" (UniqueName: \"kubernetes.io/projected/63329d01-3db3-448a-b22c-53a0d58f8497-kube-api-access-j95dt\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.033609 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.033778 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/63329d01-3db3-448a-b22c-53a0d58f8497-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.033963 4989 scope.go:117] "RemoveContainer" containerID="ca2377b7cd7f73d36c389bf39d5a5b793d4492f8b024e80813557074b081fb4c" Oct 06 09:04:40 crc kubenswrapper[4989]: E1006 09:04:40.044854 4989 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 06 09:04:40 crc kubenswrapper[4989]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-06T09:04:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 06 09:04:40 crc kubenswrapper[4989]: /etc/init.d/functions: line 589: 372 Alarm clock "$@" Oct 06 09:04:40 crc kubenswrapper[4989]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-rzncg" message=< Oct 06 09:04:40 crc kubenswrapper[4989]: Exiting ovn-controller (1) [FAILED] Oct 06 09:04:40 crc kubenswrapper[4989]: Killing ovn-controller (1) [ OK ] Oct 06 09:04:40 crc kubenswrapper[4989]: 2025-10-06T09:04:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 06 09:04:40 crc kubenswrapper[4989]: /etc/init.d/functions: line 589: 372 Alarm clock "$@" Oct 06 09:04:40 crc kubenswrapper[4989]: > Oct 06 09:04:40 crc kubenswrapper[4989]: E1006 09:04:40.044885 4989 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 06 09:04:40 crc kubenswrapper[4989]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-06T09:04:37Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 06 09:04:40 crc kubenswrapper[4989]: /etc/init.d/functions: line 589: 372 Alarm clock "$@" Oct 06 09:04:40 crc kubenswrapper[4989]: > pod="openstack/ovn-controller-rzncg" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" containerID="cri-o://23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.044918 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-rzncg" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" containerID="cri-o://23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3" gracePeriod=27 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.047641 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rzncg" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" probeResult="failure" output="" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.061095 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data" (OuterVolumeSpecName: "config-data") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.066068 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.073274 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.073629 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-central-agent" containerID="cri-o://1113ed9589562e5b58b800ea0d696a3d844f0b50b70b74e3d2b8d7f5c23c5eb4" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.076856 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="proxy-httpd" containerID="cri-o://a3e857f4d02909a9a5deea6f9f5ec9a459f3932e69093005ea9d3a9d28adff43" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.077005 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="sg-core" containerID="cri-o://6a8db167aeebd3ccb93770c425902d5284cd7da3659d4cafff002f345363260f" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.077047 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-notification-agent" containerID="cri-o://0e944028fd090111d4e8a713c7330ed17face3e9d21c00ba187b11ee5fe64bf4" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136126 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc42f\" (UniqueName: \"kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136575 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136633 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136688 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136710 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.136849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts\") pod \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\" (UID: \"e68a18fb-33b8-4ddf-95a7-46563d1ed77e\") " Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.137295 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.137309 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.139750 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.141066 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.141452 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerName="kube-state-metrics" containerID="cri-o://6e66f1425bf72188ce8811a749a7167ac2b29573e0f8e328426cf8a51343c25d" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.143700 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts" (OuterVolumeSpecName: "scripts") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.153056 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.164164 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f" (OuterVolumeSpecName: "kube-api-access-fc42f") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "kube-api-access-fc42f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.165965 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.208155 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "63329d01-3db3-448a-b22c-53a0d58f8497" (UID: "63329d01-3db3-448a-b22c-53a0d58f8497"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.215354 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.215561 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerName="memcached" containerID="cri-o://81180f30221310858c389627d3cef02db0be1d4733fdbc48816fac4ffa3f8a0f" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239549 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239571 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63329d01-3db3-448a-b22c-53a0d58f8497-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239581 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239590 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc42f\" (UniqueName: \"kubernetes.io/projected/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-kube-api-access-fc42f\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239598 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.239606 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.331982 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s6ttq"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.357889 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-nlwvf"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.373135 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s6ttq"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.384582 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-nlwvf"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.385441 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.416065 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.416341 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6c7ff48d4c-577qm" podUID="920b5c92-bd05-43f9-aae2-820bb7142635" containerName="keystone-api" containerID="cri-o://1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.426320 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.441645 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data" (OuterVolumeSpecName: "config-data") pod "e68a18fb-33b8-4ddf-95a7-46563d1ed77e" (UID: "e68a18fb-33b8-4ddf-95a7-46563d1ed77e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.449717 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.450724 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e68a18fb-33b8-4ddf-95a7-46563d1ed77e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.457203 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jbcbt"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.467058 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jbcbt"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.475198 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-551f-account-create-n8p8h"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.485200 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-551f-account-create-n8p8h"] Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.669196 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="galera" containerID="cri-o://888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" gracePeriod=30 Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.960210 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": dial tcp 10.217.0.199:8775: connect: connection refused" Oct 06 09:04:40 crc kubenswrapper[4989]: I1006 09:04:40.960461 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": dial tcp 10.217.0.199:8775: connect: connection refused" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.029913 4989 generic.go:334] "Generic (PLEG): container finished" podID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerID="a7ea9a251ec17f10791afe94beed6956bf60120b934e4c1b26a49ceedb4b0502" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.029981 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerDied","Data":"a7ea9a251ec17f10791afe94beed6956bf60120b934e4c1b26a49ceedb4b0502"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.031795 4989 generic.go:334] "Generic (PLEG): container finished" podID="c29eee10-1638-43c9-806c-9c956d10f31a" containerID="de97aa68267d73cae2bc02e80acd84df5900e6e32436971bd32cc722e73765e7" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.031852 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerDied","Data":"de97aa68267d73cae2bc02e80acd84df5900e6e32436971bd32cc722e73765e7"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.033400 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerID="4a55740396b636c8de2207c1b20cbc61bd80b8803c4aaa8ddc33405ea2cecaed" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.033451 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerDied","Data":"4a55740396b636c8de2207c1b20cbc61bd80b8803c4aaa8ddc33405ea2cecaed"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.033471 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d88db6f8d-sz2w2" event={"ID":"aba3aff5-7684-45c3-9bef-25e6133cfe6e","Type":"ContainerDied","Data":"3425ce4ec6dcc5718601388a7bdbf9d7dac148a51a4bf23672929d6c5676dbbb"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.033483 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3425ce4ec6dcc5718601388a7bdbf9d7dac148a51a4bf23672929d6c5676dbbb" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035691 4989 generic.go:334] "Generic (PLEG): container finished" podID="0336b1f4-a169-4395-be06-18c72c052240" containerID="a3e857f4d02909a9a5deea6f9f5ec9a459f3932e69093005ea9d3a9d28adff43" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035716 4989 generic.go:334] "Generic (PLEG): container finished" podID="0336b1f4-a169-4395-be06-18c72c052240" containerID="6a8db167aeebd3ccb93770c425902d5284cd7da3659d4cafff002f345363260f" exitCode=2 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035724 4989 generic.go:334] "Generic (PLEG): container finished" podID="0336b1f4-a169-4395-be06-18c72c052240" containerID="1113ed9589562e5b58b800ea0d696a3d844f0b50b70b74e3d2b8d7f5c23c5eb4" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerDied","Data":"a3e857f4d02909a9a5deea6f9f5ec9a459f3932e69093005ea9d3a9d28adff43"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035791 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerDied","Data":"6a8db167aeebd3ccb93770c425902d5284cd7da3659d4cafff002f345363260f"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.035803 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerDied","Data":"1113ed9589562e5b58b800ea0d696a3d844f0b50b70b74e3d2b8d7f5c23c5eb4"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.042728 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerStarted","Data":"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.042938 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" containerID="cri-o://db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c" gracePeriod=30 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.043255 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" containerID="cri-o://1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf" gracePeriod=30 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.044142 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.044252 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.061266 4989 generic.go:334] "Generic (PLEG): container finished" podID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerID="6e66f1425bf72188ce8811a749a7167ac2b29573e0f8e328426cf8a51343c25d" exitCode=2 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.061387 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c5263b9-2c57-4f18-86b6-864cce8af85c","Type":"ContainerDied","Data":"6e66f1425bf72188ce8811a749a7167ac2b29573e0f8e328426cf8a51343c25d"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.061417 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c5263b9-2c57-4f18-86b6-864cce8af85c","Type":"ContainerDied","Data":"8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.061498 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8089dd3c01cd4cd139a8fe2b4324374c8315576e0b564d3f3f32568aef9aec59" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.064053 4989 generic.go:334] "Generic (PLEG): container finished" podID="37d4951c-2719-4670-823e-f75163b3f472" containerID="1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e" exitCode=143 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.064148 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerDied","Data":"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.067707 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rzncg_6b5f07d2-e2c4-406e-96e3-88e24d624a58/ovn-controller/0.log" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.067767 4989 generic.go:334] "Generic (PLEG): container finished" podID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerID="23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3" exitCode=143 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.067853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg" event={"ID":"6b5f07d2-e2c4-406e-96e3-88e24d624a58","Type":"ContainerDied","Data":"23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.067879 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rzncg" event={"ID":"6b5f07d2-e2c4-406e-96e3-88e24d624a58","Type":"ContainerDied","Data":"e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.067890 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5854de3e285cc3d57ae0c54f928cb2a10830bb2d3760ca6315edbed70e2d169" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.078426 4989 generic.go:334] "Generic (PLEG): container finished" podID="74245807-f752-4209-a489-d3e746b5e1fb" containerID="9e3f96ebea5bce42b75defb3a323c9cc21c5adae1bd4c054f3fd804d7a885721" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.078531 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerDied","Data":"9e3f96ebea5bce42b75defb3a323c9cc21c5adae1bd4c054f3fd804d7a885721"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.080993 4989 generic.go:334] "Generic (PLEG): container finished" podID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerID="81180f30221310858c389627d3cef02db0be1d4733fdbc48816fac4ffa3f8a0f" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.081137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c674246-9c5c-4c8c-8d0b-360305a30818","Type":"ContainerDied","Data":"81180f30221310858c389627d3cef02db0be1d4733fdbc48816fac4ffa3f8a0f"} Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.084570 4989 generic.go:334] "Generic (PLEG): container finished" podID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerID="41c3a3ea88eda96a1475512601b4880062d4b09816817a60f531e6bc6a058912" exitCode=0 Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.084858 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.084971 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerDied","Data":"41c3a3ea88eda96a1475512601b4880062d4b09816817a60f531e6bc6a058912"} Oct 06 09:04:41 crc kubenswrapper[4989]: E1006 09:04:41.405023 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 06 09:04:41 crc kubenswrapper[4989]: E1006 09:04:41.411695 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 06 09:04:41 crc kubenswrapper[4989]: E1006 09:04:41.415339 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 06 09:04:41 crc kubenswrapper[4989]: E1006 09:04:41.415404 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="galera" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.509623 4989 scope.go:117] "RemoveContainer" containerID="d527d1e974bd70387c4d59349474b57da4a2251238da24621c65f48e2aba78db" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.529230 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-795dfbf48b-x8t2t" podStartSLOduration=7.52920897 podStartE2EDuration="7.52920897s" podCreationTimestamp="2025-10-06 09:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:04:41.070156881 +0000 UTC m=+1531.860182461" watchObservedRunningTime="2025-10-06 09:04:41.52920897 +0000 UTC m=+1532.319234550" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.529805 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.549931 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-bmrx4"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.564592 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.571680 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-spqbp"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.611088 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.633394 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.651270 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.664136 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.664710 4989 scope.go:117] "RemoveContainer" containerID="24bd4b9578823a45ed78764215744ad5b434e9e908686d6d508220f295123fc5" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.678335 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.697995 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmdn9\" (UniqueName: \"kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9\") pod \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.698067 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle\") pod \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.698183 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs\") pod \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.698396 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data\") pod \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.698558 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs\") pod \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\" (UID: \"50fc7e8a-a4c4-43ca-b672-7d8f63e25738\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.724114 4989 scope.go:117] "RemoveContainer" containerID="4368459492d84638518335c2551dd210d155262a4a5fb3137e085d2b4dab20cc" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.728914 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.739320 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9" (OuterVolumeSpecName: "kube-api-access-mmdn9") pod "50fc7e8a-a4c4-43ca-b672-7d8f63e25738" (UID: "50fc7e8a-a4c4-43ca-b672-7d8f63e25738"). InnerVolumeSpecName "kube-api-access-mmdn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.750103 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.761364 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-64d9fcfdcc-xfhh5"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.793771 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.800754 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.801438 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.801552 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.801798 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.801842 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802110 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802233 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l26ml\" (UniqueName: \"kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802270 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802299 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config\") pod \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\" (UID: \"9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8\") " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.802804 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.803745 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.803644 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.804461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.804645 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.805032 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.805148 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.805253 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmdn9\" (UniqueName: \"kubernetes.io/projected/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-kube-api-access-mmdn9\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.818141 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets" (OuterVolumeSpecName: "secrets") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.820027 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml" (OuterVolumeSpecName: "kube-api-access-l26ml") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "kube-api-access-l26ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.845983 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.860463 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": dial tcp 10.217.0.195:3000: connect: connection refused" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.909489 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.909864 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.909882 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l26ml\" (UniqueName: \"kubernetes.io/projected/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-kube-api-access-l26ml\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.910060 4989 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.924008 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data" (OuterVolumeSpecName: "config-data") pod "50fc7e8a-a4c4-43ca-b672-7d8f63e25738" (UID: "50fc7e8a-a4c4-43ca-b672-7d8f63e25738"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.932824 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.937149 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "50fc7e8a-a4c4-43ca-b672-7d8f63e25738" (UID: "50fc7e8a-a4c4-43ca-b672-7d8f63e25738"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.952950 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f04a1cc-7dd1-4987-bbb1-258b06fc369f" path="/var/lib/kubelet/pods/0f04a1cc-7dd1-4987-bbb1-258b06fc369f/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.953585 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec3e388-5714-48ab-86a5-01962c90587f" path="/var/lib/kubelet/pods/1ec3e388-5714-48ab-86a5-01962c90587f/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.954261 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2445aa06-1e2f-41fe-ae4d-a998f7874955" path="/var/lib/kubelet/pods/2445aa06-1e2f-41fe-ae4d-a998f7874955/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.956365 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39fb4b45-4966-4e65-b368-322d8c65e0c8" path="/var/lib/kubelet/pods/39fb4b45-4966-4e65-b368-322d8c65e0c8/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.957819 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" path="/var/lib/kubelet/pods/63329d01-3db3-448a-b22c-53a0d58f8497/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.958424 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708a669c-612f-4261-947e-21b21effd0f0" path="/var/lib/kubelet/pods/708a669c-612f-4261-947e-21b21effd0f0/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.959006 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80aa68cc-125e-4753-8c49-17f39da3e3f5" path="/var/lib/kubelet/pods/80aa68cc-125e-4753-8c49-17f39da3e3f5/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.960009 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da6f7ed0-2bf5-4fe3-bb88-23991b325dd5" path="/var/lib/kubelet/pods/da6f7ed0-2bf5-4fe3-bb88-23991b325dd5/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.960554 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" path="/var/lib/kubelet/pods/e68a18fb-33b8-4ddf-95a7-46563d1ed77e/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.961368 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff325bd0-b2a8-4af2-b48f-b8d9f5908474" path="/var/lib/kubelet/pods/ff325bd0-b2a8-4af2-b48f-b8d9f5908474/volumes" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.975768 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50fc7e8a-a4c4-43ca-b672-7d8f63e25738" (UID: "50fc7e8a-a4c4-43ca-b672-7d8f63e25738"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.975901 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" (UID: "9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.986668 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "50fc7e8a-a4c4-43ca-b672-7d8f63e25738" (UID: "50fc7e8a-a4c4-43ca-b672-7d8f63e25738"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:41 crc kubenswrapper[4989]: I1006 09:04:41.990673 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012471 4989 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012504 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012515 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012527 4989 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012539 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012550 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50fc7e8a-a4c4-43ca-b672-7d8f63e25738-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.012558 4989 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.101846 4989 generic.go:334] "Generic (PLEG): container finished" podID="c6fa584c-5e07-489b-aac3-a91c35981b06" containerID="db101d746c1d47c14e24a34bcdcce5598dc0c3496d53a9976015951e94afabdd" exitCode=1 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.116365 4989 generic.go:334] "Generic (PLEG): container finished" podID="b2a30a5a-7506-4880-9813-7aef02c9ca57" containerID="6346380824c4612ff378adb6555342d5be3f921d185c096add78abb7484f1343" exitCode=1 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.120063 4989 generic.go:334] "Generic (PLEG): container finished" podID="6f1de8f6-ff54-4487-b72b-13200573c940" containerID="db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c" exitCode=143 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.136696 4989 generic.go:334] "Generic (PLEG): container finished" podID="29980716-4f02-4c60-9116-e0c8f9615727" containerID="5b8b677a384e84a6b4ff0c391c55ad534766384cf466d58c6955870218f38ab5" exitCode=1 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.149501 4989 generic.go:334] "Generic (PLEG): container finished" podID="da55e23b-3ed8-41dc-875e-95cad75b9a19" containerID="252fa7d9806f4179c050fc2ded734084ed8d5aac68319c520fc701cd10091fb0" exitCode=1 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.202946 4989 generic.go:334] "Generic (PLEG): container finished" podID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerID="051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" exitCode=0 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.211119 4989 generic.go:334] "Generic (PLEG): container finished" podID="8a895ac5-a669-43f0-85d5-071adb1d9762" containerID="6213722258a0d116b6fb5d2fd515a9348785770aa81c7bb338b9918f93db1645" exitCode=1 Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.234782 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.235885 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348207 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf526-account-delete-nt44j" event={"ID":"c6fa584c-5e07-489b-aac3-a91c35981b06","Type":"ContainerDied","Data":"db101d746c1d47c14e24a34bcdcce5598dc0c3496d53a9976015951e94afabdd"} Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.348216 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774 is running failed: container process not found" containerID="051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348249 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c674246-9c5c-4c8c-8d0b-360305a30818","Type":"ContainerDied","Data":"afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348265 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afc6da303d0fb0fca392e475228935c8d20545645d8ec431fd40815ce72e7ce7" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348274 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c86165ce-d1cf-42cb-ace7-468740266dd0","Type":"ContainerDied","Data":"ae61dbab66d5f5da263e5dfec201cabd998f00bb8e59729d685cf7a7e5ffb8ee"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348286 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae61dbab66d5f5da263e5dfec201cabd998f00bb8e59729d685cf7a7e5ffb8ee" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348294 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef11-account-delete-sjgsl" event={"ID":"b2a30a5a-7506-4880-9813-7aef02c9ca57","Type":"ContainerDied","Data":"6346380824c4612ff378adb6555342d5be3f921d185c096add78abb7484f1343"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348306 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerDied","Data":"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348325 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c29eee10-1638-43c9-806c-9c956d10f31a","Type":"ContainerDied","Data":"2c549db93f26ee1b5fe2037813724772ee6b80e80fdbb7d7333383eba9d8bab2"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348335 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c549db93f26ee1b5fe2037813724772ee6b80e80fdbb7d7333383eba9d8bab2" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348342 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronec71-account-delete-8f8xz" event={"ID":"29980716-4f02-4c60-9116-e0c8f9615727","Type":"ContainerDied","Data":"5b8b677a384e84a6b4ff0c391c55ad534766384cf466d58c6955870218f38ab5"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348352 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceec63-account-delete-5qcmq" event={"ID":"da55e23b-3ed8-41dc-875e-95cad75b9a19","Type":"ContainerDied","Data":"252fa7d9806f4179c050fc2ded734084ed8d5aac68319c520fc701cd10091fb0"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348363 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74245807-f752-4209-a489-d3e746b5e1fb","Type":"ContainerDied","Data":"895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348373 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895e49c03a716b96a8611a8421524c1ecdd8d77d83ae0548dc9484f1d8f7580a" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348383 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"716f391a-794d-4511-9fc2-fa48a3a948cc","Type":"ContainerDied","Data":"051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348395 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"716f391a-794d-4511-9fc2-fa48a3a948cc","Type":"ContainerDied","Data":"04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348402 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04b8e68f01e6858c765e07cc0518156da4df04c109d44e01b8d6dde85c847d5f" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348413 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8c94-account-delete-nh5mr" event={"ID":"8a895ac5-a669-43f0-85d5-071adb1d9762","Type":"ContainerDied","Data":"6213722258a0d116b6fb5d2fd515a9348785770aa81c7bb338b9918f93db1645"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348425 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"75b2344f-bdb4-4854-a732-6069c50f41f2","Type":"ContainerDied","Data":"d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee"} Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.348434 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9587f37aeb1a75b406cfd2e88e6959a040aa6597b2952283252a6c5b65159ee" Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.353772 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774 is running failed: container process not found" containerID="051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.354152 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774 is running failed: container process not found" containerID="051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.354192 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerName="nova-scheduler-scheduler" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.382826 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.386502 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.390633 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.398065 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.402865 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.422360 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-rzncg_6b5f07d2-e2c4-406e-96e3-88e24d624a58/ovn-controller/0.log" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.422436 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.436288 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.446895 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.456520 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.503064 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.508459 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537104 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xfbk\" (UniqueName: \"kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537199 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537265 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537293 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs\") pod \"7c5263b9-2c57-4f18-86b6-864cce8af85c\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537331 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd6w7\" (UniqueName: \"kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537373 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle\") pod \"7c5263b9-2c57-4f18-86b6-864cce8af85c\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537436 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537460 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537477 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537521 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537550 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537571 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config\") pod \"7c5263b9-2c57-4f18-86b6-864cce8af85c\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537588 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537609 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537646 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537695 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537751 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537778 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb2ts\" (UniqueName: \"kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537813 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537850 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537882 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537898 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn\") pod \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\" (UID: \"6b5f07d2-e2c4-406e-96e3-88e24d624a58\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537918 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj2bs\" (UniqueName: \"kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs\") pod \"7c5263b9-2c57-4f18-86b6-864cce8af85c\" (UID: \"7c5263b9-2c57-4f18-86b6-864cce8af85c\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537944 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle\") pod \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\" (UID: \"aba3aff5-7684-45c3-9bef-25e6133cfe6e\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537968 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.537986 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id\") pod \"2e1dc821-e800-4146-b43c-55f73af4daf0\" (UID: \"2e1dc821-e800-4146-b43c-55f73af4daf0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.538869 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.543729 4989 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.551865 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts" (OuterVolumeSpecName: "scripts") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.553486 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk" (OuterVolumeSpecName: "kube-api-access-5xfbk") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "kube-api-access-5xfbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.553598 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs" (OuterVolumeSpecName: "kube-api-access-kj2bs") pod "7c5263b9-2c57-4f18-86b6-864cce8af85c" (UID: "7c5263b9-2c57-4f18-86b6-864cce8af85c"). InnerVolumeSpecName "kube-api-access-kj2bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.553954 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run" (OuterVolumeSpecName: "var-run") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.554832 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.554880 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs" (OuterVolumeSpecName: "logs") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.557320 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts" (OuterVolumeSpecName: "scripts") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.557409 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7" (OuterVolumeSpecName: "kube-api-access-kd6w7") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "kube-api-access-kd6w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.557468 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.561060 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs" (OuterVolumeSpecName: "logs") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.571080 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts" (OuterVolumeSpecName: "kube-api-access-wb2ts") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "kube-api-access-wb2ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.572702 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts" (OuterVolumeSpecName: "scripts") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.574142 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.613047 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645134 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs\") pod \"c86165ce-d1cf-42cb-ace7-468740266dd0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645221 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle\") pod \"c86165ce-d1cf-42cb-ace7-468740266dd0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645282 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs\") pod \"c86165ce-d1cf-42cb-ace7-468740266dd0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645331 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645776 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645817 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp\") pod \"c86165ce-d1cf-42cb-ace7-468740266dd0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645868 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645949 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645968 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94gp5\" (UniqueName: \"kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.645987 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs\") pod \"c29eee10-1638-43c9-806c-9c956d10f31a\" (UID: \"c29eee10-1638-43c9-806c-9c956d10f31a\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646011 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data\") pod \"c86165ce-d1cf-42cb-ace7-468740266dd0\" (UID: \"c86165ce-d1cf-42cb-ace7-468740266dd0\") " Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646407 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646424 4989 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646433 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd6w7\" (UniqueName: \"kubernetes.io/projected/2e1dc821-e800-4146-b43c-55f73af4daf0-kube-api-access-kd6w7\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646442 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e1dc821-e800-4146-b43c-55f73af4daf0-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646450 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646458 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aba3aff5-7684-45c3-9bef-25e6133cfe6e-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646466 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646474 4989 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6b5f07d2-e2c4-406e-96e3-88e24d624a58-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646482 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb2ts\" (UniqueName: \"kubernetes.io/projected/6b5f07d2-e2c4-406e-96e3-88e24d624a58-kube-api-access-wb2ts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646490 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b5f07d2-e2c4-406e-96e3-88e24d624a58-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646498 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj2bs\" (UniqueName: \"kubernetes.io/projected/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-api-access-kj2bs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646507 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e1dc821-e800-4146-b43c-55f73af4daf0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646516 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.646525 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xfbk\" (UniqueName: \"kubernetes.io/projected/aba3aff5-7684-45c3-9bef-25e6133cfe6e-kube-api-access-5xfbk\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.646586 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.646627 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data podName:03d954db-7dc6-4921-b260-1c189b9492c2 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:50.646613387 +0000 UTC m=+1541.436638967 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data") pod "rabbitmq-server-0" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2") : configmap "rabbitmq-config-data" not found Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.657829 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs" (OuterVolumeSpecName: "logs") pod "c86165ce-d1cf-42cb-ace7-468740266dd0" (UID: "c86165ce-d1cf-42cb-ace7-468740266dd0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.658272 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c5263b9-2c57-4f18-86b6-864cce8af85c" (UID: "7c5263b9-2c57-4f18-86b6-864cce8af85c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.658724 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs" (OuterVolumeSpecName: "logs") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.661170 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5" (OuterVolumeSpecName: "kube-api-access-94gp5") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "kube-api-access-94gp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.710589 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp" (OuterVolumeSpecName: "kube-api-access-5wnsp") pod "c86165ce-d1cf-42cb-ace7-468740266dd0" (UID: "c86165ce-d1cf-42cb-ace7-468740266dd0"). InnerVolumeSpecName "kube-api-access-5wnsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.732636 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.733817 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "7c5263b9-2c57-4f18-86b6-864cce8af85c" (UID: "7c5263b9-2c57-4f18-86b6-864cce8af85c"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: E1006 09:04:42.762371 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d4f5ac8_6f42_41e9_91a9_dbe6a0a056a8.slice/crio-07f49586c0630ffffbb85ba7b8b2ecdf18fab1d0db40d9459b306a05121a8583\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d4f5ac8_6f42_41e9_91a9_dbe6a0a056a8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50fc7e8a_a4c4_43ca_b672_7d8f63e25738.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50fc7e8a_a4c4_43ca_b672_7d8f63e25738.slice/crio-056cd8afe444128c40113c0f59d0428a85aa64ee07b6b59a9d92282bfabe65cb\": RecentStats: unable to find data in memory cache]" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762806 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c86165ce-d1cf-42cb-ace7-468740266dd0-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762825 4989 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762835 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c29eee10-1638-43c9-806c-9c956d10f31a-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762847 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wnsp\" (UniqueName: \"kubernetes.io/projected/c86165ce-d1cf-42cb-ace7-468740266dd0-kube-api-access-5wnsp\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762857 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762869 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94gp5\" (UniqueName: \"kubernetes.io/projected/c29eee10-1638-43c9-806c-9c956d10f31a-kube-api-access-94gp5\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.762910 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.770261 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data" (OuterVolumeSpecName: "config-data") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.804940 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.811751 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "7c5263b9-2c57-4f18-86b6-864cce8af85c" (UID: "7c5263b9-2c57-4f18-86b6-864cce8af85c"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.825583 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "6b5f07d2-e2c4-406e-96e3-88e24d624a58" (UID: "6b5f07d2-e2c4-406e-96e3-88e24d624a58"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.832592 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data" (OuterVolumeSpecName: "config-data") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.864759 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865675 4989 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c5263b9-2c57-4f18-86b6-864cce8af85c-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865699 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865708 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865716 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865725 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b5f07d2-e2c4-406e-96e3-88e24d624a58-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.865734 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.866360 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data" (OuterVolumeSpecName: "config-data") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.875870 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e1dc821-e800-4146-b43c-55f73af4daf0" (UID: "2e1dc821-e800-4146-b43c-55f73af4daf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.879782 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c86165ce-d1cf-42cb-ace7-468740266dd0" (UID: "c86165ce-d1cf-42cb-ace7-468740266dd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.882486 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data" (OuterVolumeSpecName: "config-data") pod "c86165ce-d1cf-42cb-ace7-468740266dd0" (UID: "c86165ce-d1cf-42cb-ace7-468740266dd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.887083 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.907744 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.915620 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c86165ce-d1cf-42cb-ace7-468740266dd0" (UID: "c86165ce-d1cf-42cb-ace7-468740266dd0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.918800 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c29eee10-1638-43c9-806c-9c956d10f31a" (UID: "c29eee10-1638-43c9-806c-9c956d10f31a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.977959 4989 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.977995 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978010 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978021 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978032 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1dc821-e800-4146-b43c-55f73af4daf0-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978043 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978054 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c29eee10-1638-43c9-806c-9c956d10f31a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.978065 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c86165ce-d1cf-42cb-ace7-468740266dd0-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.983797 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:42 crc kubenswrapper[4989]: I1006 09:04:42.994485 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "aba3aff5-7684-45c3-9bef-25e6133cfe6e" (UID: "aba3aff5-7684-45c3-9bef-25e6133cfe6e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.007430 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.080041 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.080069 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aba3aff5-7684-45c3-9bef-25e6133cfe6e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.080093 4989 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.080170 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data podName:2559de32-42b0-4be1-b8ea-f42383d892a3 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:51.080151741 +0000 UTC m=+1541.870177321 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data") pod "rabbitmq-cell1-server-0" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3") : configmap "rabbitmq-cell1-config-data" not found Oct 06 09:04:43 crc kubenswrapper[4989]: W1006 09:04:43.105326 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e39a37f_a1e6_41da_988b_e7eda3d6f257.slice/crio-1ae25d3bab884c00d4703d725ffb575ac9fd63d2ea33b45e5a3930ca7c3a3600 WatchSource:0}: Error finding container 1ae25d3bab884c00d4703d725ffb575ac9fd63d2ea33b45e5a3930ca7c3a3600: Status 404 returned error can't find the container with id 1ae25d3bab884c00d4703d725ffb575ac9fd63d2ea33b45e5a3930ca7c3a3600 Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.120719 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.128360 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.135597 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.154897 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.162621 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.166921 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181641 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181689 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dp84\" (UniqueName: \"kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84\") pod \"c6fa584c-5e07-489b-aac3-a91c35981b06\" (UID: \"c6fa584c-5e07-489b-aac3-a91c35981b06\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181746 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs\") pod \"4c674246-9c5c-4c8c-8d0b-360305a30818\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181790 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181823 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181847 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcc9x\" (UniqueName: \"kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x\") pod \"716f391a-794d-4511-9fc2-fa48a3a948cc\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181890 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data\") pod \"4c674246-9c5c-4c8c-8d0b-360305a30818\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181911 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181941 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181971 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.181992 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182014 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182034 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data\") pod \"716f391a-794d-4511-9fc2-fa48a3a948cc\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182065 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182086 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnf2q\" (UniqueName: \"kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q\") pod \"4c674246-9c5c-4c8c-8d0b-360305a30818\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182118 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle\") pod \"4c674246-9c5c-4c8c-8d0b-360305a30818\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182137 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config\") pod \"4c674246-9c5c-4c8c-8d0b-360305a30818\" (UID: \"4c674246-9c5c-4c8c-8d0b-360305a30818\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182164 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182194 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk6v9\" (UniqueName: \"kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182268 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle\") pod \"716f391a-794d-4511-9fc2-fa48a3a948cc\" (UID: \"716f391a-794d-4511-9fc2-fa48a3a948cc\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts\") pod \"75b2344f-bdb4-4854-a732-6069c50f41f2\" (UID: \"75b2344f-bdb4-4854-a732-6069c50f41f2\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182311 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxx2g\" (UniqueName: \"kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g\") pod \"b2a30a5a-7506-4880-9813-7aef02c9ca57\" (UID: \"b2a30a5a-7506-4880-9813-7aef02c9ca57\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182329 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182352 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77fnz\" (UniqueName: \"kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182383 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.182421 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs\") pod \"74245807-f752-4209-a489-d3e746b5e1fb\" (UID: \"74245807-f752-4209-a489-d3e746b5e1fb\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.183195 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs" (OuterVolumeSpecName: "logs") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.187595 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.188017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.188414 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data" (OuterVolumeSpecName: "config-data") pod "4c674246-9c5c-4c8c-8d0b-360305a30818" (UID: "4c674246-9c5c-4c8c-8d0b-360305a30818"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.188748 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs" (OuterVolumeSpecName: "logs") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.190018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.190751 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x" (OuterVolumeSpecName: "kube-api-access-xcc9x") pod "716f391a-794d-4511-9fc2-fa48a3a948cc" (UID: "716f391a-794d-4511-9fc2-fa48a3a948cc"). InnerVolumeSpecName "kube-api-access-xcc9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.191461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4c674246-9c5c-4c8c-8d0b-360305a30818" (UID: "4c674246-9c5c-4c8c-8d0b-360305a30818"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.193615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q" (OuterVolumeSpecName: "kube-api-access-jnf2q") pod "4c674246-9c5c-4c8c-8d0b-360305a30818" (UID: "4c674246-9c5c-4c8c-8d0b-360305a30818"). InnerVolumeSpecName "kube-api-access-jnf2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.196145 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9" (OuterVolumeSpecName: "kube-api-access-fk6v9") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "kube-api-access-fk6v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.199946 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.201291 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.207765 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84" (OuterVolumeSpecName: "kube-api-access-4dp84") pod "c6fa584c-5e07-489b-aac3-a91c35981b06" (UID: "c6fa584c-5e07-489b-aac3-a91c35981b06"). InnerVolumeSpecName "kube-api-access-4dp84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.208542 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz" (OuterVolumeSpecName: "kube-api-access-77fnz") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "kube-api-access-77fnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.210178 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.210982 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts" (OuterVolumeSpecName: "scripts") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.211142 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts" (OuterVolumeSpecName: "scripts") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.212976 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.225253 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g" (OuterVolumeSpecName: "kube-api-access-bxx2g") pod "b2a30a5a-7506-4880-9813-7aef02c9ca57" (UID: "b2a30a5a-7506-4880-9813-7aef02c9ca57"). InnerVolumeSpecName "kube-api-access-bxx2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.253027 4989 generic.go:334] "Generic (PLEG): container finished" podID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerID="888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" exitCode=0 Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.253089 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerDied","Data":"888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.259443 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceec63-account-delete-5qcmq" event={"ID":"da55e23b-3ed8-41dc-875e-95cad75b9a19","Type":"ContainerDied","Data":"7ba3bd59092f163e1a214bdb9280ce107e2e2ee6424baf9c75cd83b5c7fcd188"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.259508 4989 scope.go:117] "RemoveContainer" containerID="252fa7d9806f4179c050fc2ded734084ed8d5aac68319c520fc701cd10091fb0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.259474 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceec63-account-delete-5qcmq" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.267491 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerStarted","Data":"1ae25d3bab884c00d4703d725ffb575ac9fd63d2ea33b45e5a3930ca7c3a3600"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.274117 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderf526-account-delete-nt44j" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.274146 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderf526-account-delete-nt44j" event={"ID":"c6fa584c-5e07-489b-aac3-a91c35981b06","Type":"ContainerDied","Data":"0861bb01ae8435110462fe291366e108fc5b9d6d89ecf0d5b87fdf43199716b0"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.283556 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8c94-account-delete-nh5mr" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.283634 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8c94-account-delete-nh5mr" event={"ID":"8a895ac5-a669-43f0-85d5-071adb1d9762","Type":"ContainerDied","Data":"780521d1c60ffd8a7a82e20039a146e6d64b70dcc354c248fd0125e83ba58888"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.283821 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf7vs\" (UniqueName: \"kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs\") pod \"8a895ac5-a669-43f0-85d5-071adb1d9762\" (UID: \"8a895ac5-a669-43f0-85d5-071adb1d9762\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.283951 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg7cl\" (UniqueName: \"kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl\") pod \"29980716-4f02-4c60-9116-e0c8f9615727\" (UID: \"29980716-4f02-4c60-9116-e0c8f9615727\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.283992 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9td6\" (UniqueName: \"kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6\") pod \"da55e23b-3ed8-41dc-875e-95cad75b9a19\" (UID: \"da55e23b-3ed8-41dc-875e-95cad75b9a19\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287190 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxx2g\" (UniqueName: \"kubernetes.io/projected/b2a30a5a-7506-4880-9813-7aef02c9ca57-kube-api-access-bxx2g\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287219 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287230 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287241 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77fnz\" (UniqueName: \"kubernetes.io/projected/74245807-f752-4209-a489-d3e746b5e1fb-kube-api-access-77fnz\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287251 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287261 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dp84\" (UniqueName: \"kubernetes.io/projected/c6fa584c-5e07-489b-aac3-a91c35981b06-kube-api-access-4dp84\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287284 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287296 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcc9x\" (UniqueName: \"kubernetes.io/projected/716f391a-794d-4511-9fc2-fa48a3a948cc-kube-api-access-xcc9x\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287307 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287322 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287335 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74245807-f752-4209-a489-d3e746b5e1fb-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287345 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287356 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/75b2344f-bdb4-4854-a732-6069c50f41f2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287367 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnf2q\" (UniqueName: \"kubernetes.io/projected/4c674246-9c5c-4c8c-8d0b-360305a30818-kube-api-access-jnf2q\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287377 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c674246-9c5c-4c8c-8d0b-360305a30818-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.287389 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk6v9\" (UniqueName: \"kubernetes.io/projected/75b2344f-bdb4-4854-a732-6069c50f41f2-kube-api-access-fk6v9\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.290056 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ef11-account-delete-sjgsl" event={"ID":"b2a30a5a-7506-4880-9813-7aef02c9ca57","Type":"ContainerDied","Data":"3708f56e2bd61189da05b826d96da515af7aba3ef8b112e5c9c170019698e9d6"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.290245 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ef11-account-delete-sjgsl" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.293594 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.301146 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronec71-account-delete-8f8xz" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.301162 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronec71-account-delete-8f8xz" event={"ID":"29980716-4f02-4c60-9116-e0c8f9615727","Type":"ContainerDied","Data":"17987688491a57d4a92caa2a16b302757802609966c2e371a7024efac9c878a1"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.303689 4989 generic.go:334] "Generic (PLEG): container finished" podID="4efffd9e-87ec-4569-be26-03091469d765" containerID="aa79f2c4a47bfd9cd6e90a260e7452c6eaabae169b34d4f1882b65c8d80ca5ac" exitCode=0 Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.304077 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.304579 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerDied","Data":"aa79f2c4a47bfd9cd6e90a260e7452c6eaabae169b34d4f1882b65c8d80ca5ac"} Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.304692 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rzncg" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305218 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305211 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305375 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d88db6f8d-sz2w2" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305437 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305438 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305461 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305404 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.305636 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.333741 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl" (OuterVolumeSpecName: "kube-api-access-gg7cl") pod "29980716-4f02-4c60-9116-e0c8f9615727" (UID: "29980716-4f02-4c60-9116-e0c8f9615727"). InnerVolumeSpecName "kube-api-access-gg7cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.335616 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs" (OuterVolumeSpecName: "kube-api-access-mf7vs") pod "8a895ac5-a669-43f0-85d5-071adb1d9762" (UID: "8a895ac5-a669-43f0-85d5-071adb1d9762"). InnerVolumeSpecName "kube-api-access-mf7vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.337622 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6" (OuterVolumeSpecName: "kube-api-access-z9td6") pod "da55e23b-3ed8-41dc-875e-95cad75b9a19" (UID: "da55e23b-3ed8-41dc-875e-95cad75b9a19"). InnerVolumeSpecName "kube-api-access-z9td6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.344422 4989 scope.go:117] "RemoveContainer" containerID="db101d746c1d47c14e24a34bcdcce5598dc0c3496d53a9976015951e94afabdd" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.378700 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.388512 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderf526-account-delete-nt44j"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.388585 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.388625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.388712 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389568 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389618 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bp25\" (UniqueName: \"kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389768 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.389913 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default\") pod \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\" (UID: \"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f\") " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.390633 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf7vs\" (UniqueName: \"kubernetes.io/projected/8a895ac5-a669-43f0-85d5-071adb1d9762-kube-api-access-mf7vs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.390683 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg7cl\" (UniqueName: \"kubernetes.io/projected/29980716-4f02-4c60-9116-e0c8f9615727-kube-api-access-gg7cl\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.390694 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9td6\" (UniqueName: \"kubernetes.io/projected/da55e23b-3ed8-41dc-875e-95cad75b9a19-kube-api-access-z9td6\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.390702 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.391194 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.391924 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.391962 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.408610 4989 scope.go:117] "RemoveContainer" containerID="6213722258a0d116b6fb5d2fd515a9348785770aa81c7bb338b9918f93db1645" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.412480 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data" (OuterVolumeSpecName: "config-data") pod "716f391a-794d-4511-9fc2-fa48a3a948cc" (UID: "716f391a-794d-4511-9fc2-fa48a3a948cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.416264 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.419139 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25" (OuterVolumeSpecName: "kube-api-access-4bp25") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "kube-api-access-4bp25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.422180 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets" (OuterVolumeSpecName: "secrets") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.422227 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0ef11-account-delete-sjgsl"] Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.452692 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.459859 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.460238 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.463639 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.464382 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerName="nova-cell1-conductor-conductor" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.474329 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.488321 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494116 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494139 4989 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494168 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494180 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bp25\" (UniqueName: \"kubernetes.io/projected/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-kube-api-access-4bp25\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494190 4989 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-secrets\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494198 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.494206 4989 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.495963 4989 scope.go:117] "RemoveContainer" containerID="6346380824c4612ff378adb6555342d5be3f921d185c096add78abb7484f1343" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.499044 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.508527 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.515762 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.525322 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.534850 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.541867 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "716f391a-794d-4511-9fc2-fa48a3a948cc" (UID: "716f391a-794d-4511-9fc2-fa48a3a948cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.548373 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c674246-9c5c-4c8c-8d0b-360305a30818" (UID: "4c674246-9c5c-4c8c-8d0b-360305a30818"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.549749 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rzncg"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.569141 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.573542 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.577187 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.588467 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.590360 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.590525 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5d88db6f8d-sz2w2"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599838 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599865 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599878 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/716f391a-794d-4511-9fc2-fa48a3a948cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599888 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599897 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.599909 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.603248 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.603767 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.605501 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.619154 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.628044 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "4c674246-9c5c-4c8c-8d0b-360305a30818" (UID: "4c674246-9c5c-4c8c-8d0b-360305a30818"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.628958 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.629063 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" (UID: "894c9ae7-7bc4-49e3-9c22-e4602ea00f4f"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.632172 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data" (OuterVolumeSpecName: "config-data") pod "75b2344f-bdb4-4854-a732-6069c50f41f2" (UID: "75b2344f-bdb4-4854-a732-6069c50f41f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.638969 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data" (OuterVolumeSpecName: "config-data") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.641826 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "74245807-f752-4209-a489-d3e746b5e1fb" (UID: "74245807-f752-4209-a489-d3e746b5e1fb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700475 4989 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c674246-9c5c-4c8c-8d0b-360305a30818-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700512 4989 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700522 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700532 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700540 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700548 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/74245807-f752-4209-a489-d3e746b5e1fb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700556 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.700567 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75b2344f-bdb4-4854-a732-6069c50f41f2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.910004 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.910038 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.910047 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.910058 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:43 crc kubenswrapper[4989]: E1006 09:04:43.910102 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:04:51.91008853 +0000 UTC m=+1542.700114110 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.952868 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" path="/var/lib/kubelet/pods/2e1dc821-e800-4146-b43c-55f73af4daf0/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.953845 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" path="/var/lib/kubelet/pods/50fc7e8a-a4c4-43ca-b672-7d8f63e25738/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.954464 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" path="/var/lib/kubelet/pods/6b5f07d2-e2c4-406e-96e3-88e24d624a58/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.955802 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" path="/var/lib/kubelet/pods/7c5263b9-2c57-4f18-86b6-864cce8af85c/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.956507 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" path="/var/lib/kubelet/pods/9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.957197 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" path="/var/lib/kubelet/pods/aba3aff5-7684-45c3-9bef-25e6133cfe6e/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.958297 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a30a5a-7506-4880-9813-7aef02c9ca57" path="/var/lib/kubelet/pods/b2a30a5a-7506-4880-9813-7aef02c9ca57/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.958889 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" path="/var/lib/kubelet/pods/c29eee10-1638-43c9-806c-9c956d10f31a/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.959452 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6fa584c-5e07-489b-aac3-a91c35981b06" path="/var/lib/kubelet/pods/c6fa584c-5e07-489b-aac3-a91c35981b06/volumes" Oct 06 09:04:43 crc kubenswrapper[4989]: I1006 09:04:43.960573 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" path="/var/lib/kubelet/pods/c86165ce-d1cf-42cb-ace7-468740266dd0/volumes" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.020183 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.021514 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.022835 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.022867 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" containerName="nova-cell0-conductor-conductor" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.098183 4989 scope.go:117] "RemoveContainer" containerID="5b8b677a384e84a6b4ff0c391c55ad534766384cf466d58c6955870218f38ab5" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.109821 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.114549 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.123636 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glanceec63-account-delete-5qcmq"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.126367 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.126735 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-64d9fcfdcc-xfhh5" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": dial tcp 10.217.0.165:8080: i/o timeout" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.137584 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.192836 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.193298 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.199736 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronec71-account-delete-8f8xz"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.208999 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.213318 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican8c94-account-delete-nh5mr"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214182 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214256 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214503 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214526 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214543 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214592 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lx6d\" (UniqueName: \"kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.214611 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle\") pod \"4efffd9e-87ec-4569-be26-03091469d765\" (UID: \"4efffd9e-87ec-4569-be26-03091469d765\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.222505 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.224351 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs" (OuterVolumeSpecName: "logs") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.224419 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.225239 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d" (OuterVolumeSpecName: "kube-api-access-8lx6d") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "kube-api-access-8lx6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.225592 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.239689 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.250160 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.260714 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.266739 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.268330 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.276154 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.280912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data" (OuterVolumeSpecName: "config-data") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.285581 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.291541 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.294038 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4efffd9e-87ec-4569-be26-03091469d765" (UID: "4efffd9e-87ec-4569-be26-03091469d765"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.297244 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317000 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317048 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317078 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317116 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnn66\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317175 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317196 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317222 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317245 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317278 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317297 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317321 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317345 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317377 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317405 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317426 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317447 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317470 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317591 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317615 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317635 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.317993 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318039 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf\") pod \"2559de32-42b0-4be1-b8ea-f42383d892a3\" (UID: \"2559de32-42b0-4be1-b8ea-f42383d892a3\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318063 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljp2w\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318091 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf\") pod \"03d954db-7dc6-4921-b260-1c189b9492c2\" (UID: \"03d954db-7dc6-4921-b260-1c189b9492c2\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318361 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lx6d\" (UniqueName: \"kubernetes.io/projected/4efffd9e-87ec-4569-be26-03091469d765-kube-api-access-8lx6d\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318383 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318395 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318405 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318415 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4efffd9e-87ec-4569-be26-03091469d765-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318425 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318436 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4efffd9e-87ec-4569-be26-03091469d765-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.318922 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.319346 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.321149 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.321901 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.322450 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.323449 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66" (OuterVolumeSpecName: "kube-api-access-qnn66") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "kube-api-access-qnn66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.324800 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.325066 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.325140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.325216 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326295 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326479 4989 generic.go:334] "Generic (PLEG): container finished" podID="920b5c92-bd05-43f9-aae2-820bb7142635" containerID="1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee" exitCode=0 Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7ff48d4c-577qm" event={"ID":"920b5c92-bd05-43f9-aae2-820bb7142635","Type":"ContainerDied","Data":"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326606 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7ff48d4c-577qm" event={"ID":"920b5c92-bd05-43f9-aae2-820bb7142635","Type":"ContainerDied","Data":"2028c691029687175d5e45716e016814d0107edae5d8330358f6401cd71e286a"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326627 4989 scope.go:117] "RemoveContainer" containerID="1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.326781 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7ff48d4c-577qm" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.327443 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info" (OuterVolumeSpecName: "pod-info") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.327828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.331290 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"894c9ae7-7bc4-49e3-9c22-e4602ea00f4f","Type":"ContainerDied","Data":"f5a8d3c2ce7e605df9caf12641cfdcf6f7ad14ac049fd896947dc9ee69cd638a"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.331410 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.331550 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.333645 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info" (OuterVolumeSpecName: "pod-info") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.335395 4989 generic.go:334] "Generic (PLEG): container finished" podID="03d954db-7dc6-4921-b260-1c189b9492c2" containerID="97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032" exitCode=0 Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.335461 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerDied","Data":"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.335505 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"03d954db-7dc6-4921-b260-1c189b9492c2","Type":"ContainerDied","Data":"9a93c9894dd0374e1b8060a5f0e19a9624a479c9f00dbaa15ad2c78f5b8a5f93"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.335770 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.336968 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57c669566-xd94v" event={"ID":"4efffd9e-87ec-4569-be26-03091469d765","Type":"ContainerDied","Data":"57d1c5d3c6d2439adf9bbdac16253bc88c648586f0f6d19ed85db5c690ee6b36"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.337036 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57c669566-xd94v" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.338058 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w" (OuterVolumeSpecName: "kube-api-access-ljp2w") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "kube-api-access-ljp2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.341269 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.345125 4989 generic.go:334] "Generic (PLEG): container finished" podID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerID="5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7" exitCode=0 Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.345303 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerDied","Data":"5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.349997 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data" (OuterVolumeSpecName: "config-data") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.351093 4989 generic.go:334] "Generic (PLEG): container finished" podID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerID="e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca" exitCode=0 Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.351170 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerDied","Data":"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.351211 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2559de32-42b0-4be1-b8ea-f42383d892a3","Type":"ContainerDied","Data":"a6b77edea39c60d3ee36b124df90fe871d17d6826162373fdff14438b1068075"} Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.351308 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.354671 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data" (OuterVolumeSpecName: "config-data") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.358881 4989 scope.go:117] "RemoveContainer" containerID="1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.359500 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee\": container with ID starting with 1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee not found: ID does not exist" containerID="1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.359537 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee"} err="failed to get container status \"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee\": rpc error: code = NotFound desc = could not find container \"1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee\": container with ID starting with 1038099246d0b54452cd46aef8f33dc271d9a93e00e01cb5debea262b49287ee not found: ID does not exist" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.359559 4989 scope.go:117] "RemoveContainer" containerID="888bf91351c79d8231fa4b23597130d4134af2482ab54c9e97d6dc6644e3daf0" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.383644 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.385564 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.386047 4989 scope.go:117] "RemoveContainer" containerID="6d1aa99512d591f42008e3253eb070ca752e10603a8f70402dc5c4cf6d4c9eaa" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.395073 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.406741 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf" (OuterVolumeSpecName: "server-conf") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.409261 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.414161 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57c669566-xd94v"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.417217 4989 scope.go:117] "RemoveContainer" containerID="97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419411 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419480 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419560 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419643 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf95p\" (UniqueName: \"kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.419742 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs\") pod \"920b5c92-bd05-43f9-aae2-820bb7142635\" (UID: \"920b5c92-bd05-43f9-aae2-820bb7142635\") " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420116 4989 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420133 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420142 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnn66\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-kube-api-access-qnn66\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420151 4989 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2559de32-42b0-4be1-b8ea-f42383d892a3-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420171 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420179 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420188 4989 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/03d954db-7dc6-4921-b260-1c189b9492c2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420196 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420205 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420214 4989 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/03d954db-7dc6-4921-b260-1c189b9492c2-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420223 4989 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2559de32-42b0-4be1-b8ea-f42383d892a3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420231 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420243 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420253 4989 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420261 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420271 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420279 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420287 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420295 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420304 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljp2w\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-kube-api-access-ljp2w\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.420311 4989 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/03d954db-7dc6-4921-b260-1c189b9492c2-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.436746 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p" (OuterVolumeSpecName: "kube-api-access-xf95p") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "kube-api-access-xf95p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.437771 4989 scope.go:117] "RemoveContainer" containerID="a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.442978 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts" (OuterVolumeSpecName: "scripts") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.443945 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf" (OuterVolumeSpecName: "server-conf") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.447825 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.452811 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.456987 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data" (OuterVolumeSpecName: "config-data") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.463158 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.464430 4989 scope.go:117] "RemoveContainer" containerID="97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.464986 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032\": container with ID starting with 97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032 not found: ID does not exist" containerID="97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.465029 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032"} err="failed to get container status \"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032\": rpc error: code = NotFound desc = could not find container \"97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032\": container with ID starting with 97ca1a04206ce93dae9f02d987fb61bd22e2518b80354259f9ecbd5a418e9032 not found: ID does not exist" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.465056 4989 scope.go:117] "RemoveContainer" containerID="a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.465532 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b\": container with ID starting with a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b not found: ID does not exist" containerID="a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.465566 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b"} err="failed to get container status \"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b\": rpc error: code = NotFound desc = could not find container \"a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b\": container with ID starting with a0d751f51faaafaa766ff63c4b07cbcf21a82896b32ff7f32f253027d00f6d8b not found: ID does not exist" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.465592 4989 scope.go:117] "RemoveContainer" containerID="aa79f2c4a47bfd9cd6e90a260e7452c6eaabae169b34d4f1882b65c8d80ca5ac" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.476207 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "03d954db-7dc6-4921-b260-1c189b9492c2" (UID: "03d954db-7dc6-4921-b260-1c189b9492c2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.476968 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2559de32-42b0-4be1-b8ea-f42383d892a3" (UID: "2559de32-42b0-4be1-b8ea-f42383d892a3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.486537 4989 scope.go:117] "RemoveContainer" containerID="c0e526eeb3fc63bca07ba53b274454c122438ef8d646f58978c54082302aad38" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.490911 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.507191 4989 scope.go:117] "RemoveContainer" containerID="e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522740 4989 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2559de32-42b0-4be1-b8ea-f42383d892a3-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522778 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522791 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522804 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/03d954db-7dc6-4921-b260-1c189b9492c2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522815 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522826 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522836 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2559de32-42b0-4be1-b8ea-f42383d892a3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522846 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522857 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf95p\" (UniqueName: \"kubernetes.io/projected/920b5c92-bd05-43f9-aae2-820bb7142635-kube-api-access-xf95p\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.522868 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.524266 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "920b5c92-bd05-43f9-aae2-820bb7142635" (UID: "920b5c92-bd05-43f9-aae2-820bb7142635"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.624927 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/920b5c92-bd05-43f9-aae2-820bb7142635-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.677505 4989 scope.go:117] "RemoveContainer" containerID="081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.725887 4989 scope.go:117] "RemoveContainer" containerID="e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.731297 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca\": container with ID starting with e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca not found: ID does not exist" containerID="e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.731376 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca"} err="failed to get container status \"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca\": rpc error: code = NotFound desc = could not find container \"e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca\": container with ID starting with e82415ae2b4e7f1c8536b77b876a9afeb2ea0ee967d4826a1135974074bc00ca not found: ID does not exist" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.731405 4989 scope.go:117] "RemoveContainer" containerID="081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.732845 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4\": container with ID starting with 081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4 not found: ID does not exist" containerID="081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.732895 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4"} err="failed to get container status \"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4\": rpc error: code = NotFound desc = could not find container \"081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4\": container with ID starting with 081df7cc21957a726abce7aa3167ad1f5b954e2ab4c609d533797748c9ec88a4 not found: ID does not exist" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.763227 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.772392 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.774617 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.780147 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.780198 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.807484 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.808029 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.808407 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.808594 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.808998 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.809910 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6c7ff48d4c-577qm"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.812058 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.818053 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:44 crc kubenswrapper[4989]: E1006 09:04:44.818131 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.819306 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.829388 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.837883 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:04:44 crc kubenswrapper[4989]: I1006 09:04:44.844065 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.196916 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b485137c-1d6e-4a7a-b8db-83da01affc1a/ovn-northd/0.log" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.197185 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.239851 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.239927 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.239980 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.240045 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.240133 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.240174 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htm69\" (UniqueName: \"kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.240200 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.241250 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.241458 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config" (OuterVolumeSpecName: "config") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.241609 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts" (OuterVolumeSpecName: "scripts") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.246453 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69" (OuterVolumeSpecName: "kube-api-access-htm69") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "kube-api-access-htm69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.268522 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.324571 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.334574 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.348170 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle\") pod \"f62910dd-7c4b-43ee-bac2-c1562c633e34\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.353745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m44gr\" (UniqueName: \"kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr\") pod \"f62910dd-7c4b-43ee-bac2-c1562c633e34\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.369164 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.369436 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data\") pod \"f62910dd-7c4b-43ee-bac2-c1562c633e34\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.369769 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom\") pod \"f62910dd-7c4b-43ee-bac2-c1562c633e34\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.369797 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs\") pod \"f62910dd-7c4b-43ee-bac2-c1562c633e34\" (UID: \"f62910dd-7c4b-43ee-bac2-c1562c633e34\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.369848 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") pod \"b485137c-1d6e-4a7a-b8db-83da01affc1a\" (UID: \"b485137c-1d6e-4a7a-b8db-83da01affc1a\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370355 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370367 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370375 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htm69\" (UniqueName: \"kubernetes.io/projected/b485137c-1d6e-4a7a-b8db-83da01affc1a-kube-api-access-htm69\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370385 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370395 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b485137c-1d6e-4a7a-b8db-83da01affc1a-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370402 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: W1006 09:04:45.370452 4989 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b485137c-1d6e-4a7a-b8db-83da01affc1a/volumes/kubernetes.io~secret/metrics-certs-tls-certs Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370460 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b485137c-1d6e-4a7a-b8db-83da01affc1a" (UID: "b485137c-1d6e-4a7a-b8db-83da01affc1a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.370494 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs" (OuterVolumeSpecName: "logs") pod "f62910dd-7c4b-43ee-bac2-c1562c633e34" (UID: "f62910dd-7c4b-43ee-bac2-c1562c633e34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.379828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f62910dd-7c4b-43ee-bac2-c1562c633e34" (UID: "f62910dd-7c4b-43ee-bac2-c1562c633e34"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.379877 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f62910dd-7c4b-43ee-bac2-c1562c633e34" (UID: "f62910dd-7c4b-43ee-bac2-c1562c633e34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.379972 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr" (OuterVolumeSpecName: "kube-api-access-m44gr") pod "f62910dd-7c4b-43ee-bac2-c1562c633e34" (UID: "f62910dd-7c4b-43ee-bac2-c1562c633e34"). InnerVolumeSpecName "kube-api-access-m44gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.385874 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerID="ca7e527853be7d729e32e29acc9093bf1307493ccd1757e678492f80c2ca1c5e" exitCode=0 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.385927 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerDied","Data":"ca7e527853be7d729e32e29acc9093bf1307493ccd1757e678492f80c2ca1c5e"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.393045 4989 generic.go:334] "Generic (PLEG): container finished" podID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerID="a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" exitCode=0 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.393126 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73314917-dc1f-470c-b2d6-932ebcd443b8","Type":"ContainerDied","Data":"a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.394543 4989 generic.go:334] "Generic (PLEG): container finished" podID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerID="79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db" exitCode=0 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.394579 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerDied","Data":"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.394595 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" event={"ID":"f62910dd-7c4b-43ee-bac2-c1562c633e34","Type":"ContainerDied","Data":"5fd13a48c0ac4ff5bca90fd8daf5fdb73f10b10fcc260ff67cf9e738467accd7"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.394612 4989 scope.go:117] "RemoveContainer" containerID="79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.394720 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9c89c9ff-5cnfn" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.406453 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b485137c-1d6e-4a7a-b8db-83da01affc1a/ovn-northd/0.log" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.406491 4989 generic.go:334] "Generic (PLEG): container finished" podID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" exitCode=139 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.406589 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.407781 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerDied","Data":"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.407849 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b485137c-1d6e-4a7a-b8db-83da01affc1a","Type":"ContainerDied","Data":"2a942f4481e91b5f1414b133e3b415ffe137c9225e1a7712198ba31b70f75e61"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.410873 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data" (OuterVolumeSpecName: "config-data") pod "f62910dd-7c4b-43ee-bac2-c1562c633e34" (UID: "f62910dd-7c4b-43ee-bac2-c1562c633e34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.420897 4989 generic.go:334] "Generic (PLEG): container finished" podID="0336b1f4-a169-4395-be06-18c72c052240" containerID="0e944028fd090111d4e8a713c7330ed17face3e9d21c00ba187b11ee5fe64bf4" exitCode=0 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.420952 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerDied","Data":"0e944028fd090111d4e8a713c7330ed17face3e9d21c00ba187b11ee5fe64bf4"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.422332 4989 generic.go:334] "Generic (PLEG): container finished" podID="5a17496d-00da-47b7-a178-5274afa8ac00" containerID="275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" exitCode=0 Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.422362 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5a17496d-00da-47b7-a178-5274afa8ac00","Type":"ContainerDied","Data":"275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb"} Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.432947 4989 scope.go:117] "RemoveContainer" containerID="206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.470022 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471554 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m44gr\" (UniqueName: \"kubernetes.io/projected/f62910dd-7c4b-43ee-bac2-c1562c633e34-kube-api-access-m44gr\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471576 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471587 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471595 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62910dd-7c4b-43ee-bac2-c1562c633e34-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471605 4989 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b485137c-1d6e-4a7a-b8db-83da01affc1a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.471613 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62910dd-7c4b-43ee-bac2-c1562c633e34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.476621 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.482417 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.487771 4989 scope.go:117] "RemoveContainer" containerID="79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db" Oct 06 09:04:45 crc kubenswrapper[4989]: E1006 09:04:45.488133 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db\": container with ID starting with 79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db not found: ID does not exist" containerID="79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.488165 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db"} err="failed to get container status \"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db\": rpc error: code = NotFound desc = could not find container \"79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db\": container with ID starting with 79ec2a4e2e878ad71f6ab090f9651946a3935885c0d473a88c742636e38565db not found: ID does not exist" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.488189 4989 scope.go:117] "RemoveContainer" containerID="206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3" Oct 06 09:04:45 crc kubenswrapper[4989]: E1006 09:04:45.488678 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3\": container with ID starting with 206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3 not found: ID does not exist" containerID="206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.488708 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3"} err="failed to get container status \"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3\": rpc error: code = NotFound desc = could not find container \"206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3\": container with ID starting with 206b4c4c86243bc0a94f392d87a82978e2aaca88ebadbc26ec24a78a1f7b6bc3 not found: ID does not exist" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.488725 4989 scope.go:117] "RemoveContainer" containerID="5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.489953 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.527357 4989 scope.go:117] "RemoveContainer" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.573954 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data\") pod \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.574786 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575172 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs\") pod \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575120 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575293 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575368 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle\") pod \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575462 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575498 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom\") pod \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575532 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575558 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bb5v\" (UniqueName: \"kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v\") pod \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\" (UID: \"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575608 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575701 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575755 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp84m\" (UniqueName: \"kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575787 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs\") pod \"0336b1f4-a169-4395-be06-18c72c052240\" (UID: \"0336b1f4-a169-4395-be06-18c72c052240\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.575977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs" (OuterVolumeSpecName: "logs") pod "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" (UID: "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.576240 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.576251 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.578850 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.584154 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" (UID: "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.587959 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v" (OuterVolumeSpecName: "kube-api-access-6bb5v") pod "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" (UID: "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2"). InnerVolumeSpecName "kube-api-access-6bb5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.600984 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m" (OuterVolumeSpecName: "kube-api-access-fp84m") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "kube-api-access-fp84m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.610492 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts" (OuterVolumeSpecName: "scripts") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.646353 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" (UID: "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.664912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.665764 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data" (OuterVolumeSpecName: "config-data") pod "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" (UID: "0a41e6f6-46e1-424d-ab5e-c5ad60022ee2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678669 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp84m\" (UniqueName: \"kubernetes.io/projected/0336b1f4-a169-4395-be06-18c72c052240-kube-api-access-fp84m\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678691 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678700 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678709 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0336b1f4-a169-4395-be06-18c72c052240-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678740 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678748 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678757 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.678765 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bb5v\" (UniqueName: \"kubernetes.io/projected/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2-kube-api-access-6bb5v\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.689897 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.733940 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.763086 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data" (OuterVolumeSpecName: "config-data") pod "0336b1f4-a169-4395-be06-18c72c052240" (UID: "0336b1f4-a169-4395-be06-18c72c052240"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.771281 4989 scope.go:117] "RemoveContainer" containerID="5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d" Oct 06 09:04:45 crc kubenswrapper[4989]: E1006 09:04:45.771818 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d\": container with ID starting with 5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d not found: ID does not exist" containerID="5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.771890 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d"} err="failed to get container status \"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d\": rpc error: code = NotFound desc = could not find container \"5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d\": container with ID starting with 5ba15e96c783a064da4db03784584bb2e5a4fe8c42056e2e2f690da3f0232c2d not found: ID does not exist" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.771916 4989 scope.go:117] "RemoveContainer" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" Oct 06 09:04:45 crc kubenswrapper[4989]: E1006 09:04:45.772325 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be\": container with ID starting with b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be not found: ID does not exist" containerID="b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.772366 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be"} err="failed to get container status \"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be\": rpc error: code = NotFound desc = could not find container \"b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be\": container with ID starting with b507e39205e7a705842677db2e75ae3014f8df576fcf60dfd7a067683f45e9be not found: ID does not exist" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.780623 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.780691 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.780706 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336b1f4-a169-4395-be06-18c72c052240-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.785532 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.798268 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.801184 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.805745 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7c9c89c9ff-5cnfn"] Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882160 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8w6w\" (UniqueName: \"kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w\") pod \"5a17496d-00da-47b7-a178-5274afa8ac00\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882214 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plr2k\" (UniqueName: \"kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k\") pod \"73314917-dc1f-470c-b2d6-932ebcd443b8\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882246 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data\") pod \"5a17496d-00da-47b7-a178-5274afa8ac00\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882268 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data\") pod \"73314917-dc1f-470c-b2d6-932ebcd443b8\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882356 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle\") pod \"5a17496d-00da-47b7-a178-5274afa8ac00\" (UID: \"5a17496d-00da-47b7-a178-5274afa8ac00\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.882419 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle\") pod \"73314917-dc1f-470c-b2d6-932ebcd443b8\" (UID: \"73314917-dc1f-470c-b2d6-932ebcd443b8\") " Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.886281 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k" (OuterVolumeSpecName: "kube-api-access-plr2k") pod "73314917-dc1f-470c-b2d6-932ebcd443b8" (UID: "73314917-dc1f-470c-b2d6-932ebcd443b8"). InnerVolumeSpecName "kube-api-access-plr2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.886853 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w" (OuterVolumeSpecName: "kube-api-access-n8w6w") pod "5a17496d-00da-47b7-a178-5274afa8ac00" (UID: "5a17496d-00da-47b7-a178-5274afa8ac00"). InnerVolumeSpecName "kube-api-access-n8w6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.902900 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data" (OuterVolumeSpecName: "config-data") pod "73314917-dc1f-470c-b2d6-932ebcd443b8" (UID: "73314917-dc1f-470c-b2d6-932ebcd443b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.906437 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73314917-dc1f-470c-b2d6-932ebcd443b8" (UID: "73314917-dc1f-470c-b2d6-932ebcd443b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.906841 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a17496d-00da-47b7-a178-5274afa8ac00" (UID: "5a17496d-00da-47b7-a178-5274afa8ac00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.910837 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data" (OuterVolumeSpecName: "config-data") pod "5a17496d-00da-47b7-a178-5274afa8ac00" (UID: "5a17496d-00da-47b7-a178-5274afa8ac00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.946469 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" path="/var/lib/kubelet/pods/03d954db-7dc6-4921-b260-1c189b9492c2/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.947807 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" path="/var/lib/kubelet/pods/2559de32-42b0-4be1-b8ea-f42383d892a3/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.949004 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29980716-4f02-4c60-9116-e0c8f9615727" path="/var/lib/kubelet/pods/29980716-4f02-4c60-9116-e0c8f9615727/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.949669 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" path="/var/lib/kubelet/pods/4c674246-9c5c-4c8c-8d0b-360305a30818/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.950285 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4efffd9e-87ec-4569-be26-03091469d765" path="/var/lib/kubelet/pods/4efffd9e-87ec-4569-be26-03091469d765/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.951482 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" path="/var/lib/kubelet/pods/716f391a-794d-4511-9fc2-fa48a3a948cc/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.952294 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74245807-f752-4209-a489-d3e746b5e1fb" path="/var/lib/kubelet/pods/74245807-f752-4209-a489-d3e746b5e1fb/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.953391 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" path="/var/lib/kubelet/pods/75b2344f-bdb4-4854-a732-6069c50f41f2/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.954772 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" path="/var/lib/kubelet/pods/894c9ae7-7bc4-49e3-9c22-e4602ea00f4f/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.955812 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a895ac5-a669-43f0-85d5-071adb1d9762" path="/var/lib/kubelet/pods/8a895ac5-a669-43f0-85d5-071adb1d9762/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.956714 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920b5c92-bd05-43f9-aae2-820bb7142635" path="/var/lib/kubelet/pods/920b5c92-bd05-43f9-aae2-820bb7142635/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.957984 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" path="/var/lib/kubelet/pods/b485137c-1d6e-4a7a-b8db-83da01affc1a/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.958581 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da55e23b-3ed8-41dc-875e-95cad75b9a19" path="/var/lib/kubelet/pods/da55e23b-3ed8-41dc-875e-95cad75b9a19/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.959270 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" path="/var/lib/kubelet/pods/f62910dd-7c4b-43ee-bac2-c1562c633e34/volumes" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987057 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987097 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8w6w\" (UniqueName: \"kubernetes.io/projected/5a17496d-00da-47b7-a178-5274afa8ac00-kube-api-access-n8w6w\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987108 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plr2k\" (UniqueName: \"kubernetes.io/projected/73314917-dc1f-470c-b2d6-932ebcd443b8-kube-api-access-plr2k\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987116 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987125 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73314917-dc1f-470c-b2d6-932ebcd443b8-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:45 crc kubenswrapper[4989]: I1006 09:04:45.987133 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a17496d-00da-47b7-a178-5274afa8ac00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.436818 4989 generic.go:334] "Generic (PLEG): container finished" podID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerID="84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d" exitCode=0 Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.436865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerDied","Data":"84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d"} Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.441156 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0336b1f4-a169-4395-be06-18c72c052240","Type":"ContainerDied","Data":"12b835224b464f10981293d881a362bbc1d3a5512530f5acba148e8ca14df5e3"} Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.441183 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.441209 4989 scope.go:117] "RemoveContainer" containerID="a3e857f4d02909a9a5deea6f9f5ec9a459f3932e69093005ea9d3a9d28adff43" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.445518 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5a17496d-00da-47b7-a178-5274afa8ac00","Type":"ContainerDied","Data":"1911e063ec7efbf835478a5f0ace21b2f1ac61866b1fec81f151df77f0beb741"} Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.445620 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.448103 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"73314917-dc1f-470c-b2d6-932ebcd443b8","Type":"ContainerDied","Data":"f3f823bab6661fa109dab69c29d53e8bb7313b585270c503620ce968978c078f"} Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.448193 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.456190 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" event={"ID":"0a41e6f6-46e1-424d-ab5e-c5ad60022ee2","Type":"ContainerDied","Data":"8c2e64b1115e348874a19be0c8f6bb19bc04442309cc5a7bd10dc4c828b3c765"} Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.456252 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84fbfd87f8-2zrs8" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.467306 4989 scope.go:117] "RemoveContainer" containerID="6a8db167aeebd3ccb93770c425902d5284cd7da3659d4cafff002f345363260f" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.489409 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.501942 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.504352 4989 scope.go:117] "RemoveContainer" containerID="0e944028fd090111d4e8a713c7330ed17face3e9d21c00ba187b11ee5fe64bf4" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.510056 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.558132 4989 scope.go:117] "RemoveContainer" containerID="1113ed9589562e5b58b800ea0d696a3d844f0b50b70b74e3d2b8d7f5c23c5eb4" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.558172 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.565852 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.574155 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.576876 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.580319 4989 scope.go:117] "RemoveContainer" containerID="275760ebf0df73ff6a7b03bbdf781fe78d6e72f40e3b33bacd1add098a502deb" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.584495 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-84fbfd87f8-2zrs8"] Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.604325 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.193:6080/vnc_lite.html\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.604515 4989 scope.go:117] "RemoveContainer" containerID="a627a8399e3c086968ceb500a961383c60739e7f8df3f5089aec7cefe5699dbe" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.626979 4989 scope.go:117] "RemoveContainer" containerID="ca7e527853be7d729e32e29acc9093bf1307493ccd1757e678492f80c2ca1c5e" Oct 06 09:04:46 crc kubenswrapper[4989]: I1006 09:04:46.647086 4989 scope.go:117] "RemoveContainer" containerID="fac01ec109fb091c49f98d9c013903a2e4a8fcce4fac1a9f3b69848da9287623" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.432352 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.169:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.478830 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerStarted","Data":"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de"} Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.509007 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7fnl7" podStartSLOduration=7.953779434 podStartE2EDuration="10.508986864s" podCreationTimestamp="2025-10-06 09:04:37 +0000 UTC" firstStartedPulling="2025-10-06 09:04:44.35877301 +0000 UTC m=+1535.148798590" lastFinishedPulling="2025-10-06 09:04:46.91398044 +0000 UTC m=+1537.704006020" observedRunningTime="2025-10-06 09:04:47.506006998 +0000 UTC m=+1538.296032588" watchObservedRunningTime="2025-10-06 09:04:47.508986864 +0000 UTC m=+1538.299012444" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.944187 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0336b1f4-a169-4395-be06-18c72c052240" path="/var/lib/kubelet/pods/0336b1f4-a169-4395-be06-18c72c052240/volumes" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.945014 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" path="/var/lib/kubelet/pods/0a41e6f6-46e1-424d-ab5e-c5ad60022ee2/volumes" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.945604 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" path="/var/lib/kubelet/pods/5a17496d-00da-47b7-a178-5274afa8ac00/volumes" Oct 06 09:04:47 crc kubenswrapper[4989]: I1006 09:04:47.946623 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" path="/var/lib/kubelet/pods/73314917-dc1f-470c-b2d6-932ebcd443b8/volumes" Oct 06 09:04:48 crc kubenswrapper[4989]: I1006 09:04:48.007438 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.103:11211: i/o timeout" Oct 06 09:04:48 crc kubenswrapper[4989]: I1006 09:04:48.995716 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c669566-xd94v" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:48 crc kubenswrapper[4989]: I1006 09:04:48.995797 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57c669566-xd94v" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": context deadline exceeded" Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.804730 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.805284 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.805496 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.805525 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.806326 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.811012 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.812406 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:49 crc kubenswrapper[4989]: E1006 09:04:49.812478 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:04:50 crc kubenswrapper[4989]: I1006 09:04:50.584356 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:50 crc kubenswrapper[4989]: I1006 09:04:50.591092 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:52 crc kubenswrapper[4989]: E1006 09:04:52.004072 4989 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Oct 06 09:04:52 crc kubenswrapper[4989]: E1006 09:04:52.004128 4989 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Oct 06 09:04:52 crc kubenswrapper[4989]: E1006 09:04:52.004138 4989 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 06 09:04:52 crc kubenswrapper[4989]: E1006 09:04:52.004151 4989 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:52 crc kubenswrapper[4989]: E1006 09:04:52.004202 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift podName:aaf6911f-dbaf-432c-9103-71f58a087a12 nodeName:}" failed. No retries permitted until 2025-10-06 09:05:08.004184708 +0000 UTC m=+1558.794210288 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift") pod "swift-storage-0" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Oct 06 09:04:52 crc kubenswrapper[4989]: I1006 09:04:52.387794 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:52 crc kubenswrapper[4989]: I1006 09:04:52.387849 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:52 crc kubenswrapper[4989]: I1006 09:04:52.437251 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:52 crc kubenswrapper[4989]: I1006 09:04:52.568495 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:52 crc kubenswrapper[4989]: I1006 09:04:52.672266 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:54 crc kubenswrapper[4989]: I1006 09:04:54.543428 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7fnl7" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="registry-server" containerID="cri-o://baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de" gracePeriod=2 Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.805343 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.806783 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.809753 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.811044 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.811126 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.811193 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.819710 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:54 crc kubenswrapper[4989]: E1006 09:04:54.819840 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.051564 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.150256 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities\") pod \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.150313 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8vrz\" (UniqueName: \"kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz\") pod \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.150531 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content\") pod \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\" (UID: \"8e39a37f-a1e6-41da-988b-e7eda3d6f257\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.152299 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities" (OuterVolumeSpecName: "utilities") pod "8e39a37f-a1e6-41da-988b-e7eda3d6f257" (UID: "8e39a37f-a1e6-41da-988b-e7eda3d6f257"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.155594 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz" (OuterVolumeSpecName: "kube-api-access-m8vrz") pod "8e39a37f-a1e6-41da-988b-e7eda3d6f257" (UID: "8e39a37f-a1e6-41da-988b-e7eda3d6f257"). InnerVolumeSpecName "kube-api-access-m8vrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.169311 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252398 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252491 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252593 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252747 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhcwg\" (UniqueName: \"kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252878 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252919 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.252958 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs\") pod \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\" (UID: \"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0\") " Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.253363 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.253380 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8vrz\" (UniqueName: \"kubernetes.io/projected/8e39a37f-a1e6-41da-988b-e7eda3d6f257-kube-api-access-m8vrz\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.253938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e39a37f-a1e6-41da-988b-e7eda3d6f257" (UID: "8e39a37f-a1e6-41da-988b-e7eda3d6f257"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.255953 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg" (OuterVolumeSpecName: "kube-api-access-lhcwg") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "kube-api-access-lhcwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.257243 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.289078 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.289347 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.291518 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config" (OuterVolumeSpecName: "config") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.293280 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.315535 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" (UID: "241e6b97-f00e-4efe-8ce3-3a3b8da1aef0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354471 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354513 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354524 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354532 4989 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354540 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354548 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e39a37f-a1e6-41da-988b-e7eda3d6f257-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354557 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhcwg\" (UniqueName: \"kubernetes.io/projected/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-kube-api-access-lhcwg\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.354565 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.558989 4989 generic.go:334] "Generic (PLEG): container finished" podID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerID="4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875" exitCode=0 Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.559094 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerDied","Data":"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875"} Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.559135 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557b4f9dd9-blscg" event={"ID":"241e6b97-f00e-4efe-8ce3-3a3b8da1aef0","Type":"ContainerDied","Data":"05c1fe635c1b79fed0a61153e67b08bf58566ccc0cd5071305ba2f6160b7a052"} Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.559163 4989 scope.go:117] "RemoveContainer" containerID="85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.559531 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557b4f9dd9-blscg" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.564229 4989 generic.go:334] "Generic (PLEG): container finished" podID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerID="baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de" exitCode=0 Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.564285 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerDied","Data":"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de"} Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.564315 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7fnl7" event={"ID":"8e39a37f-a1e6-41da-988b-e7eda3d6f257","Type":"ContainerDied","Data":"1ae25d3bab884c00d4703d725ffb575ac9fd63d2ea33b45e5a3930ca7c3a3600"} Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.564289 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7fnl7" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.598837 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.604844 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.616428 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.618250 4989 scope.go:117] "RemoveContainer" containerID="4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.629154 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7fnl7"] Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.635373 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.646592 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-557b4f9dd9-blscg"] Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.649712 4989 scope.go:117] "RemoveContainer" containerID="85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2" Oct 06 09:04:55 crc kubenswrapper[4989]: E1006 09:04:55.650321 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2\": container with ID starting with 85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2 not found: ID does not exist" containerID="85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.650367 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2"} err="failed to get container status \"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2\": rpc error: code = NotFound desc = could not find container \"85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2\": container with ID starting with 85399b075f54ba6ece96770be23b089078e19c9c3cec15e500b182fdc99aced2 not found: ID does not exist" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.650395 4989 scope.go:117] "RemoveContainer" containerID="4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875" Oct 06 09:04:55 crc kubenswrapper[4989]: E1006 09:04:55.650746 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875\": container with ID starting with 4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875 not found: ID does not exist" containerID="4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.650789 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875"} err="failed to get container status \"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875\": rpc error: code = NotFound desc = could not find container \"4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875\": container with ID starting with 4787c6a6e8e3078c4a693698faa5d415eec9fe3b11cef703ee9c9163e636f875 not found: ID does not exist" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.650814 4989 scope.go:117] "RemoveContainer" containerID="baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.672014 4989 scope.go:117] "RemoveContainer" containerID="84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.693895 4989 scope.go:117] "RemoveContainer" containerID="5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.740610 4989 scope.go:117] "RemoveContainer" containerID="baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de" Oct 06 09:04:55 crc kubenswrapper[4989]: E1006 09:04:55.741324 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de\": container with ID starting with baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de not found: ID does not exist" containerID="baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.741356 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de"} err="failed to get container status \"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de\": rpc error: code = NotFound desc = could not find container \"baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de\": container with ID starting with baa18fea33a5c21e34a310a491d2dc031a305db16465684097109c8aa8a675de not found: ID does not exist" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.741380 4989 scope.go:117] "RemoveContainer" containerID="84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d" Oct 06 09:04:55 crc kubenswrapper[4989]: E1006 09:04:55.741620 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d\": container with ID starting with 84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d not found: ID does not exist" containerID="84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.741643 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d"} err="failed to get container status \"84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d\": rpc error: code = NotFound desc = could not find container \"84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d\": container with ID starting with 84a6df34c1ac87be7baa22d71074c6ac69b6c0663506438b3b27d277af292a3d not found: ID does not exist" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.741668 4989 scope.go:117] "RemoveContainer" containerID="5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7" Oct 06 09:04:55 crc kubenswrapper[4989]: E1006 09:04:55.741870 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7\": container with ID starting with 5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7 not found: ID does not exist" containerID="5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.741899 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7"} err="failed to get container status \"5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7\": rpc error: code = NotFound desc = could not find container \"5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7\": container with ID starting with 5166c52dbb32cd6fcd80a3c6121cfcc9e5e4ee43eeb946972c75ed9e9f627ac7 not found: ID does not exist" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.945084 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" path="/var/lib/kubelet/pods/241e6b97-f00e-4efe-8ce3-3a3b8da1aef0/volumes" Oct 06 09:04:55 crc kubenswrapper[4989]: I1006 09:04:55.945677 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" path="/var/lib/kubelet/pods/8e39a37f-a1e6-41da-988b-e7eda3d6f257/volumes" Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.804408 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.805631 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.806216 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.806262 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.806432 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.807545 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.809325 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:04:59 crc kubenswrapper[4989]: E1006 09:04:59.809374 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:05:00 crc kubenswrapper[4989]: I1006 09:05:00.603837 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:00 crc kubenswrapper[4989]: I1006 09:05:00.609245 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.804485 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.805332 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.805433 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.805871 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.805901 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.806884 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.808507 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 06 09:05:04 crc kubenswrapper[4989]: E1006 09:05:04.808552 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-tmq59" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:05:05 crc kubenswrapper[4989]: I1006 09:05:05.649865 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:05 crc kubenswrapper[4989]: I1006 09:05:05.650387 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.677771 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tmq59_29637231-ec41-4e57-af1f-18046cc06cf6/ovs-vswitchd/0.log" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.679060 4989 generic.go:334] "Generic (PLEG): container finished" podID="29637231-ec41-4e57-af1f-18046cc06cf6" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" exitCode=137 Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.679109 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerDied","Data":"9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825"} Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.830960 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tmq59_29637231-ec41-4e57-af1f-18046cc06cf6/ovs-vswitchd/0.log" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.833037 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.936799 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.936874 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.936933 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.936969 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937051 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937092 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkfbt\" (UniqueName: \"kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt\") pod \"29637231-ec41-4e57-af1f-18046cc06cf6\" (UID: \"29637231-ec41-4e57-af1f-18046cc06cf6\") " Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937871 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937916 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run" (OuterVolumeSpecName: "var-run") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937927 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib" (OuterVolumeSpecName: "var-lib") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.937957 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log" (OuterVolumeSpecName: "var-log") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.939383 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts" (OuterVolumeSpecName: "scripts") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:05:06 crc kubenswrapper[4989]: I1006 09:05:06.943669 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt" (OuterVolumeSpecName: "kube-api-access-rkfbt") pod "29637231-ec41-4e57-af1f-18046cc06cf6" (UID: "29637231-ec41-4e57-af1f-18046cc06cf6"). InnerVolumeSpecName "kube-api-access-rkfbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039201 4989 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039240 4989 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-run\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039369 4989 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-log\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039521 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkfbt\" (UniqueName: \"kubernetes.io/projected/29637231-ec41-4e57-af1f-18046cc06cf6-kube-api-access-rkfbt\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039575 4989 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29637231-ec41-4e57-af1f-18046cc06cf6-var-lib\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.039585 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29637231-ec41-4e57-af1f-18046cc06cf6-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.176911 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.242541 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock\") pod \"aaf6911f-dbaf-432c-9103-71f58a087a12\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.242694 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") pod \"aaf6911f-dbaf-432c-9103-71f58a087a12\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.242744 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"aaf6911f-dbaf-432c-9103-71f58a087a12\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.242814 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z5xf\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf\") pod \"aaf6911f-dbaf-432c-9103-71f58a087a12\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.242842 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache\") pod \"aaf6911f-dbaf-432c-9103-71f58a087a12\" (UID: \"aaf6911f-dbaf-432c-9103-71f58a087a12\") " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.243573 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock" (OuterVolumeSpecName: "lock") pod "aaf6911f-dbaf-432c-9103-71f58a087a12" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.243790 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache" (OuterVolumeSpecName: "cache") pod "aaf6911f-dbaf-432c-9103-71f58a087a12" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.246889 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "aaf6911f-dbaf-432c-9103-71f58a087a12" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.247026 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf" (OuterVolumeSpecName: "kube-api-access-7z5xf") pod "aaf6911f-dbaf-432c-9103-71f58a087a12" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12"). InnerVolumeSpecName "kube-api-access-7z5xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.250800 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "swift") pod "aaf6911f-dbaf-432c-9103-71f58a087a12" (UID: "aaf6911f-dbaf-432c-9103-71f58a087a12"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.344955 4989 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-lock\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.345249 4989 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.345284 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.345294 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z5xf\" (UniqueName: \"kubernetes.io/projected/aaf6911f-dbaf-432c-9103-71f58a087a12-kube-api-access-7z5xf\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.345305 4989 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aaf6911f-dbaf-432c-9103-71f58a087a12-cache\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.358982 4989 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.446878 4989 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.691915 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-tmq59_29637231-ec41-4e57-af1f-18046cc06cf6/ovs-vswitchd/0.log" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.693010 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-tmq59" event={"ID":"29637231-ec41-4e57-af1f-18046cc06cf6","Type":"ContainerDied","Data":"e37df7547070a50f2c8867346b9be9b45cff209265c630183e91bd4a1f02eeea"} Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.693067 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-tmq59" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.693074 4989 scope.go:117] "RemoveContainer" containerID="9ac6a90a089a8005f33e2773bfc69fc3b21467507675a07b2c64874cfdc2b825" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.701930 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerID="f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c" exitCode=137 Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.701979 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c"} Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.702018 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aaf6911f-dbaf-432c-9103-71f58a087a12","Type":"ContainerDied","Data":"7209a166318b3f15faf2edd793ce1ca9f607f07e72732eb9a45314021ae21935"} Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.702166 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.718768 4989 scope.go:117] "RemoveContainer" containerID="8f9704af963389dee20ee1677e7d9f4722b21dbed03612406f761cae4a2c93de" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.730799 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.741570 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-tmq59"] Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.757147 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.762559 4989 scope.go:117] "RemoveContainer" containerID="db6a98c9f7876b97505eaf206617e7f0a26dba08f6ba38b5d990e6763b2bee16" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.762914 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.800227 4989 scope.go:117] "RemoveContainer" containerID="f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.820827 4989 scope.go:117] "RemoveContainer" containerID="ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.840143 4989 scope.go:117] "RemoveContainer" containerID="8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.859972 4989 scope.go:117] "RemoveContainer" containerID="3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.875949 4989 scope.go:117] "RemoveContainer" containerID="6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.892953 4989 scope.go:117] "RemoveContainer" containerID="662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.908290 4989 scope.go:117] "RemoveContainer" containerID="0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.935323 4989 scope.go:117] "RemoveContainer" containerID="4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.949296 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" path="/var/lib/kubelet/pods/29637231-ec41-4e57-af1f-18046cc06cf6/volumes" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.950142 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" path="/var/lib/kubelet/pods/aaf6911f-dbaf-432c-9103-71f58a087a12/volumes" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.959523 4989 scope.go:117] "RemoveContainer" containerID="6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.974790 4989 scope.go:117] "RemoveContainer" containerID="02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3" Oct 06 09:05:07 crc kubenswrapper[4989]: I1006 09:05:07.989487 4989 scope.go:117] "RemoveContainer" containerID="855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.032235 4989 scope.go:117] "RemoveContainer" containerID="e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.050196 4989 scope.go:117] "RemoveContainer" containerID="e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.071818 4989 scope.go:117] "RemoveContainer" containerID="9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.090356 4989 scope.go:117] "RemoveContainer" containerID="15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.107437 4989 scope.go:117] "RemoveContainer" containerID="f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.107975 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c\": container with ID starting with f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c not found: ID does not exist" containerID="f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.108256 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c"} err="failed to get container status \"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c\": rpc error: code = NotFound desc = could not find container \"f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c\": container with ID starting with f47914613a3fadd0712c7a2d999e11cc131a116264a45a829d3a194907b3d22c not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.108322 4989 scope.go:117] "RemoveContainer" containerID="ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.108738 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636\": container with ID starting with ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636 not found: ID does not exist" containerID="ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.108778 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636"} err="failed to get container status \"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636\": rpc error: code = NotFound desc = could not find container \"ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636\": container with ID starting with ead352b22d977a3b6d83299ed7ea1c1c372c36aaf2878b6fedc3454f60654636 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.108802 4989 scope.go:117] "RemoveContainer" containerID="8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.109085 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051\": container with ID starting with 8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051 not found: ID does not exist" containerID="8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.109113 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051"} err="failed to get container status \"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051\": rpc error: code = NotFound desc = could not find container \"8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051\": container with ID starting with 8920667c11ba95afa5ee6d7e52b46837e3f5769263ad3356c7c0b26ccbab4051 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.109131 4989 scope.go:117] "RemoveContainer" containerID="3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.109762 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e\": container with ID starting with 3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e not found: ID does not exist" containerID="3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.109788 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e"} err="failed to get container status \"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e\": rpc error: code = NotFound desc = could not find container \"3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e\": container with ID starting with 3a535c8034c10f8b3a2ceded84cbdb284936af511a0b0cb2bc14866a0e24c68e not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.109803 4989 scope.go:117] "RemoveContainer" containerID="6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.110143 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91\": container with ID starting with 6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91 not found: ID does not exist" containerID="6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.110169 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91"} err="failed to get container status \"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91\": rpc error: code = NotFound desc = could not find container \"6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91\": container with ID starting with 6446dc6d63c0db4d4b9dcb8d8317f6af61ee1e13bc406d3f0681cab125a3dc91 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.110186 4989 scope.go:117] "RemoveContainer" containerID="662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.110533 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc\": container with ID starting with 662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc not found: ID does not exist" containerID="662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.110560 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc"} err="failed to get container status \"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc\": rpc error: code = NotFound desc = could not find container \"662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc\": container with ID starting with 662fd0814e87812957e5a6659328781e178d428aa54cee03785bf34195f697bc not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.110580 4989 scope.go:117] "RemoveContainer" containerID="0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.111105 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d\": container with ID starting with 0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d not found: ID does not exist" containerID="0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.111139 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d"} err="failed to get container status \"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d\": rpc error: code = NotFound desc = could not find container \"0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d\": container with ID starting with 0a5237b15074e90a96489ac46c928f228a9bed65145490a159238d2829f6308d not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.111164 4989 scope.go:117] "RemoveContainer" containerID="4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.111499 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889\": container with ID starting with 4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889 not found: ID does not exist" containerID="4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.111540 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889"} err="failed to get container status \"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889\": rpc error: code = NotFound desc = could not find container \"4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889\": container with ID starting with 4e7da57125bfc9c7e87ed2583d66eec4e4f5bf808ad255c32e09a430c3cc6889 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.111560 4989 scope.go:117] "RemoveContainer" containerID="6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.112104 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75\": container with ID starting with 6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75 not found: ID does not exist" containerID="6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.112135 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75"} err="failed to get container status \"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75\": rpc error: code = NotFound desc = could not find container \"6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75\": container with ID starting with 6c370e1b04b50208b434180a10d9bc037a884a21b06b8e19f404c7d931b0bd75 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.112153 4989 scope.go:117] "RemoveContainer" containerID="02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.113374 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3\": container with ID starting with 02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3 not found: ID does not exist" containerID="02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.113408 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3"} err="failed to get container status \"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3\": rpc error: code = NotFound desc = could not find container \"02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3\": container with ID starting with 02f08241bab255071a1af22356ef912bd03c5c55649acdf0be81d402d484e9d3 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.113436 4989 scope.go:117] "RemoveContainer" containerID="855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.113887 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64\": container with ID starting with 855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64 not found: ID does not exist" containerID="855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.113917 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64"} err="failed to get container status \"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64\": rpc error: code = NotFound desc = could not find container \"855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64\": container with ID starting with 855f15789d1fcdba8651c3a68834e98ab8507a66fb3fe6e5804dbf180c338d64 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.113934 4989 scope.go:117] "RemoveContainer" containerID="e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.114197 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751\": container with ID starting with e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751 not found: ID does not exist" containerID="e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114229 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751"} err="failed to get container status \"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751\": rpc error: code = NotFound desc = could not find container \"e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751\": container with ID starting with e78d453b9f5088b88fb5e3d9a8e62062f3b2dd69116d5a9d777c6bb8fe2c0751 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114249 4989 scope.go:117] "RemoveContainer" containerID="e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.114475 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9\": container with ID starting with e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9 not found: ID does not exist" containerID="e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114508 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9"} err="failed to get container status \"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9\": rpc error: code = NotFound desc = could not find container \"e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9\": container with ID starting with e9b7becd213f5704a43e98606d90d00d04ee1777c9a9b8cc8f0f25473f7c26b9 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114526 4989 scope.go:117] "RemoveContainer" containerID="9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.114830 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130\": container with ID starting with 9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130 not found: ID does not exist" containerID="9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114858 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130"} err="failed to get container status \"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130\": rpc error: code = NotFound desc = could not find container \"9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130\": container with ID starting with 9611f6ba65cb4e32a91654a8ed936b0203edf5924f19fa066ddf5df508e6a130 not found: ID does not exist" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.114870 4989 scope.go:117] "RemoveContainer" containerID="15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c" Oct 06 09:05:08 crc kubenswrapper[4989]: E1006 09:05:08.115136 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c\": container with ID starting with 15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c not found: ID does not exist" containerID="15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c" Oct 06 09:05:08 crc kubenswrapper[4989]: I1006 09:05:08.115174 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c"} err="failed to get container status \"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c\": rpc error: code = NotFound desc = could not find container \"15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c\": container with ID starting with 15ce6eb5b73d4d16139509bd1c00f86e151ddfa0108b09666a90ea3cdd12d45c not found: ID does not exist" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.052547 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.171396 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom\") pod \"bc313ada-f4bc-4711-838a-b809c4e29273\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.171497 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data\") pod \"bc313ada-f4bc-4711-838a-b809c4e29273\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.171581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle\") pod \"bc313ada-f4bc-4711-838a-b809c4e29273\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.171649 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs\") pod \"bc313ada-f4bc-4711-838a-b809c4e29273\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.171694 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssmsk\" (UniqueName: \"kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk\") pod \"bc313ada-f4bc-4711-838a-b809c4e29273\" (UID: \"bc313ada-f4bc-4711-838a-b809c4e29273\") " Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.172415 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs" (OuterVolumeSpecName: "logs") pod "bc313ada-f4bc-4711-838a-b809c4e29273" (UID: "bc313ada-f4bc-4711-838a-b809c4e29273"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.176124 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk" (OuterVolumeSpecName: "kube-api-access-ssmsk") pod "bc313ada-f4bc-4711-838a-b809c4e29273" (UID: "bc313ada-f4bc-4711-838a-b809c4e29273"). InnerVolumeSpecName "kube-api-access-ssmsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.177845 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bc313ada-f4bc-4711-838a-b809c4e29273" (UID: "bc313ada-f4bc-4711-838a-b809c4e29273"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.190980 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc313ada-f4bc-4711-838a-b809c4e29273" (UID: "bc313ada-f4bc-4711-838a-b809c4e29273"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.207921 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data" (OuterVolumeSpecName: "config-data") pod "bc313ada-f4bc-4711-838a-b809c4e29273" (UID: "bc313ada-f4bc-4711-838a-b809c4e29273"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.273356 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.273676 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.273742 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc313ada-f4bc-4711-838a-b809c4e29273-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.273814 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssmsk\" (UniqueName: \"kubernetes.io/projected/bc313ada-f4bc-4711-838a-b809c4e29273-kube-api-access-ssmsk\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.273877 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc313ada-f4bc-4711-838a-b809c4e29273-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.725588 4989 generic.go:334] "Generic (PLEG): container finished" podID="bc313ada-f4bc-4711-838a-b809c4e29273" containerID="66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2" exitCode=137 Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.725625 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68c46dd79-4twxt" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.725633 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerDied","Data":"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2"} Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.725675 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68c46dd79-4twxt" event={"ID":"bc313ada-f4bc-4711-838a-b809c4e29273","Type":"ContainerDied","Data":"22318f22bce93e507587fad2cc347182d920498706fced4191b2e168c632aeca"} Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.725695 4989 scope.go:117] "RemoveContainer" containerID="66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.754379 4989 scope.go:117] "RemoveContainer" containerID="7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.761818 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.770268 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-68c46dd79-4twxt"] Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.777043 4989 scope.go:117] "RemoveContainer" containerID="66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2" Oct 06 09:05:09 crc kubenswrapper[4989]: E1006 09:05:09.777409 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2\": container with ID starting with 66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2 not found: ID does not exist" containerID="66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.777449 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2"} err="failed to get container status \"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2\": rpc error: code = NotFound desc = could not find container \"66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2\": container with ID starting with 66a678b84edc0c59291edb7c77c5a2fe540dd0f3007ae08178c55afb6ac159c2 not found: ID does not exist" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.777476 4989 scope.go:117] "RemoveContainer" containerID="7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed" Oct 06 09:05:09 crc kubenswrapper[4989]: E1006 09:05:09.777935 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed\": container with ID starting with 7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed not found: ID does not exist" containerID="7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.777978 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed"} err="failed to get container status \"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed\": rpc error: code = NotFound desc = could not find container \"7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed\": container with ID starting with 7250881881a8459fc4ef5837ec3340b50ce2b7bf9c10abf93ab8f13463ea90ed not found: ID does not exist" Oct 06 09:05:09 crc kubenswrapper[4989]: I1006 09:05:09.945519 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" path="/var/lib/kubelet/pods/bc313ada-f4bc-4711-838a-b809c4e29273/volumes" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.239614 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.331407 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle\") pod \"37d4951c-2719-4670-823e-f75163b3f472\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.331484 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs\") pod \"37d4951c-2719-4670-823e-f75163b3f472\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.331514 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data\") pod \"37d4951c-2719-4670-823e-f75163b3f472\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.331557 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fmfm\" (UniqueName: \"kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm\") pod \"37d4951c-2719-4670-823e-f75163b3f472\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.331589 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom\") pod \"37d4951c-2719-4670-823e-f75163b3f472\" (UID: \"37d4951c-2719-4670-823e-f75163b3f472\") " Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.332226 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs" (OuterVolumeSpecName: "logs") pod "37d4951c-2719-4670-823e-f75163b3f472" (UID: "37d4951c-2719-4670-823e-f75163b3f472"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.332531 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37d4951c-2719-4670-823e-f75163b3f472-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.335380 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm" (OuterVolumeSpecName: "kube-api-access-9fmfm") pod "37d4951c-2719-4670-823e-f75163b3f472" (UID: "37d4951c-2719-4670-823e-f75163b3f472"). InnerVolumeSpecName "kube-api-access-9fmfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.336842 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "37d4951c-2719-4670-823e-f75163b3f472" (UID: "37d4951c-2719-4670-823e-f75163b3f472"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.350105 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37d4951c-2719-4670-823e-f75163b3f472" (UID: "37d4951c-2719-4670-823e-f75163b3f472"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.364938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data" (OuterVolumeSpecName: "config-data") pod "37d4951c-2719-4670-823e-f75163b3f472" (UID: "37d4951c-2719-4670-823e-f75163b3f472"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.434333 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.434573 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.434588 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fmfm\" (UniqueName: \"kubernetes.io/projected/37d4951c-2719-4670-823e-f75163b3f472-kube-api-access-9fmfm\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.434603 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37d4951c-2719-4670-823e-f75163b3f472-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.659891 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.659938 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.737797 4989 generic.go:334] "Generic (PLEG): container finished" podID="37d4951c-2719-4670-823e-f75163b3f472" containerID="87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3" exitCode=137 Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.737870 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.737879 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerDied","Data":"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3"} Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.737912 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-567987cf6d-htd97" event={"ID":"37d4951c-2719-4670-823e-f75163b3f472","Type":"ContainerDied","Data":"ea2ceaf34d5fb5c8c9f8618b2a32ab47ebe35c68225b35d9a38c2cc03f49fd2d"} Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.737930 4989 scope.go:117] "RemoveContainer" containerID="87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.767673 4989 scope.go:117] "RemoveContainer" containerID="1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.779995 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.785832 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-567987cf6d-htd97"] Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.786853 4989 scope.go:117] "RemoveContainer" containerID="87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3" Oct 06 09:05:10 crc kubenswrapper[4989]: E1006 09:05:10.787268 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3\": container with ID starting with 87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3 not found: ID does not exist" containerID="87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.787303 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3"} err="failed to get container status \"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3\": rpc error: code = NotFound desc = could not find container \"87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3\": container with ID starting with 87991184a39b86e9b65e8a426ed3511e824b39359351cfec306eba08987c24a3 not found: ID does not exist" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.787328 4989 scope.go:117] "RemoveContainer" containerID="1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e" Oct 06 09:05:10 crc kubenswrapper[4989]: E1006 09:05:10.787639 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e\": container with ID starting with 1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e not found: ID does not exist" containerID="1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e" Oct 06 09:05:10 crc kubenswrapper[4989]: I1006 09:05:10.787692 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e"} err="failed to get container status \"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e\": rpc error: code = NotFound desc = could not find container \"1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e\": container with ID starting with 1295ef776f4eab23d62961938c7267168dc4f45ac1efc627a91933bc72b6906e not found: ID does not exist" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.083790 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": EOF" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.090825 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795dfbf48b-x8t2t" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.203:9311/healthcheck\": EOF" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.515489 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649037 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649143 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649188 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649206 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649239 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgx8r\" (UniqueName: \"kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649288 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data\") pod \"6f1de8f6-ff54-4487-b72b-13200573c940\" (UID: \"6f1de8f6-ff54-4487-b72b-13200573c940\") " Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.649926 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs" (OuterVolumeSpecName: "logs") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.654491 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.655066 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r" (OuterVolumeSpecName: "kube-api-access-dgx8r") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "kube-api-access-dgx8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.671298 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.687212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.688131 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data" (OuterVolumeSpecName: "config-data") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.696563 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6f1de8f6-ff54-4487-b72b-13200573c940" (UID: "6f1de8f6-ff54-4487-b72b-13200573c940"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750777 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f1de8f6-ff54-4487-b72b-13200573c940-logs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750816 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750832 4989 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750843 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750867 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgx8r\" (UniqueName: \"kubernetes.io/projected/6f1de8f6-ff54-4487-b72b-13200573c940-kube-api-access-dgx8r\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750878 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.750887 4989 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f1de8f6-ff54-4487-b72b-13200573c940-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.752445 4989 generic.go:334] "Generic (PLEG): container finished" podID="6f1de8f6-ff54-4487-b72b-13200573c940" containerID="1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf" exitCode=137 Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.752514 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795dfbf48b-x8t2t" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.752513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerDied","Data":"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf"} Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.752622 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795dfbf48b-x8t2t" event={"ID":"6f1de8f6-ff54-4487-b72b-13200573c940","Type":"ContainerDied","Data":"7ec3f01a0b55542dcc99e22e3799f8d5c190928fd4f5bdccdf80b385eaafeaad"} Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.752649 4989 scope.go:117] "RemoveContainer" containerID="1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.786480 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.789777 4989 scope.go:117] "RemoveContainer" containerID="db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.793778 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-795dfbf48b-x8t2t"] Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.807729 4989 scope.go:117] "RemoveContainer" containerID="1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf" Oct 06 09:05:11 crc kubenswrapper[4989]: E1006 09:05:11.808144 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf\": container with ID starting with 1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf not found: ID does not exist" containerID="1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.808182 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf"} err="failed to get container status \"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf\": rpc error: code = NotFound desc = could not find container \"1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf\": container with ID starting with 1740d220d0b9a02edba4df92e7d65a34be1e295ca674d77300b33b1b6d0cf8cf not found: ID does not exist" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.808204 4989 scope.go:117] "RemoveContainer" containerID="db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c" Oct 06 09:05:11 crc kubenswrapper[4989]: E1006 09:05:11.808522 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c\": container with ID starting with db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c not found: ID does not exist" containerID="db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.808588 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c"} err="failed to get container status \"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c\": rpc error: code = NotFound desc = could not find container \"db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c\": container with ID starting with db847be1ec7c28f6e4ca3949165d19f8383beeaf67f1bc1a2eb8d25102ec667c not found: ID does not exist" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.944781 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d4951c-2719-4670-823e-f75163b3f472" path="/var/lib/kubelet/pods/37d4951c-2719-4670-823e-f75163b3f472/volumes" Oct 06 09:05:11 crc kubenswrapper[4989]: I1006 09:05:11.945480 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" path="/var/lib/kubelet/pods/6f1de8f6-ff54-4487-b72b-13200573c940/volumes" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.032297 4989 scope.go:117] "RemoveContainer" containerID="23469e85f2888277d7b579f1029a105117aaacecd7b90890eb0bed9955a350f3" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.065546 4989 scope.go:117] "RemoveContainer" containerID="e253d79428da6dc08fc7c77770484784d6ac2aeb7efd7e3493edb9854d9065e6" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.089691 4989 scope.go:117] "RemoveContainer" containerID="c82381046702a2641e3eb2d845ec36b56f2bc1a4b89e7f088496bca8507491e7" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.115931 4989 scope.go:117] "RemoveContainer" containerID="36be71fb07cf2fc2a215efb620b9b16bfbaddc2690904b596fe8d3c21c1cef65" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.138037 4989 scope.go:117] "RemoveContainer" containerID="a2ae3fd46babdad278c9956c887f5c124a9304e8a4b2bcc60897796e72e63572" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.158343 4989 scope.go:117] "RemoveContainer" containerID="c0e57d5d18f141af4479840759a41e612d6a1d950914d8c0d137ed2381bc7ce1" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.183446 4989 scope.go:117] "RemoveContainer" containerID="0337cf022056d72d28b84790f3e2c0d6ff357bbd527ef1b5bb1231fc471cd03f" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.210686 4989 scope.go:117] "RemoveContainer" containerID="5e27806c6c9bbfa7148ab8855e4f5365764ef0a9865db06ea4f1be2251cbccb8" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.229592 4989 scope.go:117] "RemoveContainer" containerID="50a79f09910e9fff54369d71bc6811a462f4e436d855283ce8d406341063c274" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.263353 4989 scope.go:117] "RemoveContainer" containerID="0722bd521228744da04f6113db075a5725d5c5a71bf59d2e590dba6347e4a201" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.286845 4989 scope.go:117] "RemoveContainer" containerID="81180f30221310858c389627d3cef02db0be1d4733fdbc48816fac4ffa3f8a0f" Oct 06 09:06:16 crc kubenswrapper[4989]: I1006 09:06:16.306431 4989 scope.go:117] "RemoveContainer" containerID="a2eaeb715dab747daba65704d4937d60bc6b107c498217cba8e70206fbcef182" Oct 06 09:06:33 crc kubenswrapper[4989]: I1006 09:06:33.935144 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:06:33 crc kubenswrapper[4989]: I1006 09:06:33.935747 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:07:03 crc kubenswrapper[4989]: I1006 09:07:03.935205 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:07:03 crc kubenswrapper[4989]: I1006 09:07:03.935776 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.467844 4989 scope.go:117] "RemoveContainer" containerID="94109ac5a382b61cf0a1a9382b747b9d85dd07d814f0ac516a8ddf57eba0b542" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.499856 4989 scope.go:117] "RemoveContainer" containerID="03207da3ec8bb49a944e98ed52107432f3635b4287bebfcfc5db7845bce4ce5e" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.557999 4989 scope.go:117] "RemoveContainer" containerID="9526030eb4478f8d3f9e0d4d119b074bfb7ed1ad9556aa7940abe108adaf3afc" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.593472 4989 scope.go:117] "RemoveContainer" containerID="ef0a28225fec2703d402a5d120a0cb3520c4282520d26b5cba5f5b3fce0d8cb7" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.639191 4989 scope.go:117] "RemoveContainer" containerID="bec9e0cb8bf60ca2e27812afea65e265672c3c12101dea95cbd9385d1fbfa24e" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.679126 4989 scope.go:117] "RemoveContainer" containerID="3c033b104a71382421eb5e78709f0cd6d1f8852588b2f3002b7703097aa35b09" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.698387 4989 scope.go:117] "RemoveContainer" containerID="293912f3fefab3ea52c04fc0cbc8a86445f49ef67f2bc2ec27278a34efebdaf8" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.729913 4989 scope.go:117] "RemoveContainer" containerID="4700f2988f629fa1cc64cc9b1c2d7c4c3f12b989c2563fc3915d0197974d5911" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.750499 4989 scope.go:117] "RemoveContainer" containerID="9fa32a8a48db3070a364396328866252d3262f83875879eea172b131eb0e8b9b" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.779896 4989 scope.go:117] "RemoveContainer" containerID="c2f6d460e7859022f2d7bbbb38123c418fd428a488e6bb639d0fbd444a72b562" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.799559 4989 scope.go:117] "RemoveContainer" containerID="3dc88fb3bd8ee12db2b77ecab178b7c37436701311d624fc0d00ff7ddc15a05d" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.822769 4989 scope.go:117] "RemoveContainer" containerID="4171f48766da159f34b60e071af620e6d01dc004f4bc666240c769ad99c40d7f" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.842919 4989 scope.go:117] "RemoveContainer" containerID="4bc2c7953a81e332c86eef320b2122bd2683241d8c8673b379fa2d669b4fcd77" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.860091 4989 scope.go:117] "RemoveContainer" containerID="73a1d6a1544a1d3a0f92a2eb8d39ec0d74705060013a79d77caacf592dd5e439" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.875682 4989 scope.go:117] "RemoveContainer" containerID="b27261eec2fed4866d7d50179573af96501d6417fcce76932c9276012042a4f0" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.918072 4989 scope.go:117] "RemoveContainer" containerID="47d832e9c2490821149fdcd781bb935c878b27774b65fc445c59b2365810ab2f" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.945993 4989 scope.go:117] "RemoveContainer" containerID="c6d2d6d34ef679612f601313213c70302260925b0c73a08758af27e449b56af2" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.969608 4989 scope.go:117] "RemoveContainer" containerID="035f91d38dc192d523703fa81988c0eb5125a06f1b63c99461bf3c49883ad233" Oct 06 09:07:16 crc kubenswrapper[4989]: I1006 09:07:16.994063 4989 scope.go:117] "RemoveContainer" containerID="316d6cc011d092b3ddc376cbf4d0f73db645ad8effcfef4c3215df2b5159f29d" Oct 06 09:07:33 crc kubenswrapper[4989]: I1006 09:07:33.935888 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:07:33 crc kubenswrapper[4989]: I1006 09:07:33.938112 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:07:33 crc kubenswrapper[4989]: I1006 09:07:33.951559 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:07:33 crc kubenswrapper[4989]: I1006 09:07:33.952726 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:07:33 crc kubenswrapper[4989]: I1006 09:07:33.952788 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" gracePeriod=600 Oct 06 09:07:34 crc kubenswrapper[4989]: E1006 09:07:34.075883 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:07:34 crc kubenswrapper[4989]: I1006 09:07:34.130490 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" exitCode=0 Oct 06 09:07:34 crc kubenswrapper[4989]: I1006 09:07:34.130546 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6"} Oct 06 09:07:34 crc kubenswrapper[4989]: I1006 09:07:34.131670 4989 scope.go:117] "RemoveContainer" containerID="78a43651e7ce75784a5023571ad86d1983f552ab991c3873f5b54242ed1b0aec" Oct 06 09:07:34 crc kubenswrapper[4989]: I1006 09:07:34.134312 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:07:34 crc kubenswrapper[4989]: E1006 09:07:34.135821 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:07:46 crc kubenswrapper[4989]: I1006 09:07:46.936330 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:07:46 crc kubenswrapper[4989]: E1006 09:07:46.937460 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:07:59 crc kubenswrapper[4989]: I1006 09:07:59.940917 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:07:59 crc kubenswrapper[4989]: E1006 09:07:59.941828 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:08:10 crc kubenswrapper[4989]: I1006 09:08:10.936442 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:08:10 crc kubenswrapper[4989]: E1006 09:08:10.937283 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.274830 4989 scope.go:117] "RemoveContainer" containerID="6e66f1425bf72188ce8811a749a7167ac2b29573e0f8e328426cf8a51343c25d" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.300917 4989 scope.go:117] "RemoveContainer" containerID="aef057b8eb8304f36a1d6725c1a189e2b948c62649b59415ad9742a99c463eb1" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.324044 4989 scope.go:117] "RemoveContainer" containerID="9f10809e8f9d1a0a063095ddee06d65c9b24be4d66c8b5f733b0a9a20eedc6ae" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.351754 4989 scope.go:117] "RemoveContainer" containerID="f4b25eb0db619fd2778bd141569f6d60f1ea4b0b8e5c38cb07786cb6a462723c" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.374761 4989 scope.go:117] "RemoveContainer" containerID="c8539b390134394ace4899f855e06be6f8d730be158ebdbf93df2ba73394ecaa" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.417170 4989 scope.go:117] "RemoveContainer" containerID="1099b3c6e3fd133aabba584c4cf9b090c37f9c1f70d86e0fdfad871f4f32cd21" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.433466 4989 scope.go:117] "RemoveContainer" containerID="61197413d388b49b459e9e0efe53d5c352fb7b325ef677d33c7410cb203309dc" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.456766 4989 scope.go:117] "RemoveContainer" containerID="35ffcc7392dc41bc132668ec57014e8ac2e8039a4359f8333605eabdde751c1e" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.476137 4989 scope.go:117] "RemoveContainer" containerID="4f5bd568ffa89c45154d075061eda3a39f78f1665fe2c860e6ff950b52542672" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.491064 4989 scope.go:117] "RemoveContainer" containerID="e2d916bd8ce663a1374a2102f14126cf7f39b34e74293d78b95861c7f07ea483" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.518167 4989 scope.go:117] "RemoveContainer" containerID="0bf97fec682ff1dbd5c988be1156a9e788940eb34663d369552ca96c1e405f14" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.544280 4989 scope.go:117] "RemoveContainer" containerID="25d25936563097e7fdc38f6b1fcc6b1675c765f066a2f410c419974bfe7533a9" Oct 06 09:08:17 crc kubenswrapper[4989]: I1006 09:08:17.563538 4989 scope.go:117] "RemoveContainer" containerID="4a55740396b636c8de2207c1b20cbc61bd80b8803c4aaa8ddc33405ea2cecaed" Oct 06 09:08:22 crc kubenswrapper[4989]: I1006 09:08:22.936361 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:08:22 crc kubenswrapper[4989]: E1006 09:08:22.937014 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:08:33 crc kubenswrapper[4989]: I1006 09:08:33.938530 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:08:33 crc kubenswrapper[4989]: E1006 09:08:33.941263 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:08:48 crc kubenswrapper[4989]: I1006 09:08:48.937360 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:08:48 crc kubenswrapper[4989]: E1006 09:08:48.938312 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:09:00 crc kubenswrapper[4989]: I1006 09:09:00.935620 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:09:00 crc kubenswrapper[4989]: E1006 09:09:00.936581 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:09:14 crc kubenswrapper[4989]: I1006 09:09:14.936416 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:09:14 crc kubenswrapper[4989]: E1006 09:09:14.937683 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.698161 4989 scope.go:117] "RemoveContainer" containerID="1503150658fd3c0f67b76860cf73488324464b757cbc4426f8c36b6c4913eb11" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.716396 4989 scope.go:117] "RemoveContainer" containerID="bc352a1f9f91a5b814eb79f29c22ab04be980b414e6d6507bfcec57d13be32b1" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.736064 4989 scope.go:117] "RemoveContainer" containerID="9e3f96ebea5bce42b75defb3a323c9cc21c5adae1bd4c054f3fd804d7a885721" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.767670 4989 scope.go:117] "RemoveContainer" containerID="58c5d959d29623aea9a09ac52e22bed81527dfc5edac480b7610c6595fb84cfd" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.818259 4989 scope.go:117] "RemoveContainer" containerID="43c2e169810eb04bada5f35eaaf81350cd70184b618b76a271f06ac3f1d73440" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.843233 4989 scope.go:117] "RemoveContainer" containerID="3814cda71279e66c461650e21aeab490084205cdc04eb79209f77094180bab6e" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.869722 4989 scope.go:117] "RemoveContainer" containerID="b5185fa6bcba77dca8b60346a7556b03433f6545c232af16b93dcb0b73118fd4" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.890144 4989 scope.go:117] "RemoveContainer" containerID="26863d5b0d9b2d53be9b4e56d60c8e66f73ac31b5497f62f6b98628cde8cac5d" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.916333 4989 scope.go:117] "RemoveContainer" containerID="42b5b43509a943241d2558a5a7d5a648587db28f3877b2329b996d7e657a1eea" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.933244 4989 scope.go:117] "RemoveContainer" containerID="a7ea9a251ec17f10791afe94beed6956bf60120b934e4c1b26a49ceedb4b0502" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.956321 4989 scope.go:117] "RemoveContainer" containerID="ebb72deedd186195f6b578c0c2fa80fae603ae9607847a578e8bb536f1883b67" Oct 06 09:09:17 crc kubenswrapper[4989]: I1006 09:09:17.975008 4989 scope.go:117] "RemoveContainer" containerID="16a49375a3c8f1a4186ec2a1c51cfb2a9d9d97c9799db9d7fe6d33abe94affef" Oct 06 09:09:18 crc kubenswrapper[4989]: I1006 09:09:18.015745 4989 scope.go:117] "RemoveContainer" containerID="0ba87e5984d61addcfc335a8689464e448fcc9f8d299b926cd462ad4453f562e" Oct 06 09:09:26 crc kubenswrapper[4989]: I1006 09:09:26.935967 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:09:26 crc kubenswrapper[4989]: E1006 09:09:26.936678 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:09:37 crc kubenswrapper[4989]: I1006 09:09:37.935938 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:09:37 crc kubenswrapper[4989]: E1006 09:09:37.936588 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:09:51 crc kubenswrapper[4989]: I1006 09:09:51.936206 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:09:51 crc kubenswrapper[4989]: E1006 09:09:51.936937 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:10:02 crc kubenswrapper[4989]: I1006 09:10:02.936609 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:10:02 crc kubenswrapper[4989]: E1006 09:10:02.937591 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.659399 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660384 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="swift-recon-cron" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660418 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="swift-recon-cron" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660442 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-notification-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660460 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-notification-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660492 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerName="kube-state-metrics" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660511 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerName="kube-state-metrics" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660546 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerName="nova-cell1-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660562 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerName="nova-cell1-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660583 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="setup-container" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660599 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="setup-container" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660626 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660643 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660847 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660868 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660892 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-reaper" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660909 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-reaper" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660944 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.660960 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.660986 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" containerName="nova-cell0-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661002 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" containerName="nova-cell0-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661028 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661046 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661068 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661085 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661112 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661129 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661149 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="extract-utilities" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661165 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="extract-utilities" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661192 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661210 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661235 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="mysql-bootstrap" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661252 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="mysql-bootstrap" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661277 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661293 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661319 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661336 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661362 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661378 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661410 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="openstack-network-exporter" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661428 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="openstack-network-exporter" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661450 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661466 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661498 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661517 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661537 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661552 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661577 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661594 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661617 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661633 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661649 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661701 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661722 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661738 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661759 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920b5c92-bd05-43f9-aae2-820bb7142635" containerName="keystone-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661776 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="920b5c92-bd05-43f9-aae2-820bb7142635" containerName="keystone-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661795 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661813 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661843 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a895ac5-a669-43f0-85d5-071adb1d9762" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661860 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a895ac5-a669-43f0-85d5-071adb1d9762" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661890 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661905 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661934 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="cinder-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661951 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="cinder-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.661975 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.661990 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662009 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662026 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662058 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29980716-4f02-4c60-9116-e0c8f9615727" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662074 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="29980716-4f02-4c60-9116-e0c8f9615727" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662099 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662116 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662142 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-expirer" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662160 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-expirer" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662189 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662210 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662232 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662248 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662269 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="probe" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662284 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="probe" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662315 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662332 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662358 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662374 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662402 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server-init" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662418 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server-init" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662441 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-central-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662456 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-central-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662482 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="setup-container" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662496 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="setup-container" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662513 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662525 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662546 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662558 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662573 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662585 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662598 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="extract-content" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662610 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="extract-content" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662625 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="rsync" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662636 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="rsync" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662788 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da55e23b-3ed8-41dc-875e-95cad75b9a19" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662811 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="da55e23b-3ed8-41dc-875e-95cad75b9a19" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662838 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662854 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662891 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662939 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662960 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="sg-core" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.662976 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="sg-core" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.662998 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663015 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663040 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663052 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663072 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663084 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-api" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663102 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663114 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663133 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663144 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663158 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fa584c-5e07-489b-aac3-a91c35981b06" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663170 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fa584c-5e07-489b-aac3-a91c35981b06" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663190 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="registry-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663202 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="registry-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663219 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663232 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663249 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663261 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663279 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663292 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663313 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663325 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663340 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="mysql-bootstrap" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663352 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="mysql-bootstrap" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663368 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663379 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663400 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663412 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663432 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663444 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663462 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663473 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663489 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a30a5a-7506-4880-9813-7aef02c9ca57" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663501 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a30a5a-7506-4880-9813-7aef02c9ca57" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663521 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerName="memcached" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663533 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerName="memcached" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663546 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663558 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663574 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerName="nova-scheduler-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663585 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerName="nova-scheduler-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663601 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663612 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663625 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663637 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663682 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663694 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: E1006 09:10:06.663715 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.663727 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664611 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a17496d-00da-47b7-a178-5274afa8ac00" containerName="nova-cell0-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664647 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664713 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664739 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2559de32-42b0-4be1-b8ea-f42383d892a3" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664759 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664778 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="cinder-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664811 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d4f5ac8-6f42-41e9-91a9-dbe6a0a056a8" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664841 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664859 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e39a37f-a1e6-41da-988b-e7eda3d6f257" containerName="registry-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664883 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="50fc7e8a-a4c4-43ca-b672-7d8f63e25738" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664915 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664947 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664975 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="29980716-4f02-4c60-9116-e0c8f9615727" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.664998 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665020 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665051 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="920b5c92-bd05-43f9-aae2-820bb7142635" containerName="keystone-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665075 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665089 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665121 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="ovn-northd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665139 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68a18fb-33b8-4ddf-95a7-46563d1ed77e" containerName="probe" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665161 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665212 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665239 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665268 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665296 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a30a5a-7506-4880-9813-7aef02c9ca57" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.665325 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d4951c-2719-4670-823e-f75163b3f472" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666016 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c674246-9c5c-4c8c-8d0b-360305a30818" containerName="memcached" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666056 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a41e6f6-46e1-424d-ab5e-c5ad60022ee2" containerName="barbican-keystone-listener-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666121 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-central-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666143 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666200 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-log" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666220 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="ceilometer-notification-agent" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666240 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666291 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666308 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="swift-recon-cron" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666323 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="241e6b97-f00e-4efe-8ce3-3a3b8da1aef0" containerName="neutron-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666378 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-reaper" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666397 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="73314917-dc1f-470c-b2d6-932ebcd443b8" containerName="nova-cell1-conductor-conductor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666465 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666496 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba3aff5-7684-45c3-9bef-25e6133cfe6e" containerName="placement-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666519 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c29eee10-1638-43c9-806c-9c956d10f31a" containerName="nova-api-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666544 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666569 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a895ac5-a669-43f0-85d5-071adb1d9762" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666586 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666604 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efffd9e-87ec-4569-be26-03091469d765" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666625 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="716f391a-794d-4511-9fc2-fa48a3a948cc" containerName="nova-scheduler-scheduler" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666649 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f1de8f6-ff54-4487-b72b-13200573c940" containerName="barbican-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666703 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-updater" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666733 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b485137c-1d6e-4a7a-b8db-83da01affc1a" containerName="openstack-network-exporter" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666752 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovsdb-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666782 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86165ce-d1cf-42cb-ace7-468740266dd0" containerName="nova-metadata-metadata" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666824 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5f07d2-e2c4-406e-96e3-88e24d624a58" containerName="ovn-controller" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666851 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666911 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b2344f-bdb4-4854-a732-6069c50f41f2" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666937 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc313ada-f4bc-4711-838a-b809c4e29273" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666960 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336b1f4-a169-4395-be06-18c72c052240" containerName="sg-core" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666981 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-expirer" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.666997 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="container-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667019 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1dc821-e800-4146-b43c-55f73af4daf0" containerName="cinder-api" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667047 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="account-replicator" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667072 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d954db-7dc6-4921-b260-1c189b9492c2" containerName="rabbitmq" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667088 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="63329d01-3db3-448a-b22c-53a0d58f8497" containerName="proxy-server" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667108 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="894c9ae7-7bc4-49e3-9c22-e4602ea00f4f" containerName="galera" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667133 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="object-auditor" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667155 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="da55e23b-3ed8-41dc-875e-95cad75b9a19" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667178 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5263b9-2c57-4f18-86b6-864cce8af85c" containerName="kube-state-metrics" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667197 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf6911f-dbaf-432c-9103-71f58a087a12" containerName="rsync" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667213 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62910dd-7c4b-43ee-bac2-c1562c633e34" containerName="barbican-worker" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667236 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="74245807-f752-4209-a489-d3e746b5e1fb" containerName="glance-httpd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667252 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fa584c-5e07-489b-aac3-a91c35981b06" containerName="mariadb-account-delete" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.667278 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="29637231-ec41-4e57-af1f-18046cc06cf6" containerName="ovs-vswitchd" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.669831 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.682035 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.784378 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.784561 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.784736 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.885594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.885689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.885777 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.886430 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.886545 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:06 crc kubenswrapper[4989]: I1006 09:10:06.917487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg\") pod \"community-operators-p45wp\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:07 crc kubenswrapper[4989]: I1006 09:10:07.006921 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:07 crc kubenswrapper[4989]: I1006 09:10:07.351455 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:07 crc kubenswrapper[4989]: I1006 09:10:07.495268 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerStarted","Data":"44a0a3b69b20ffbea65bfd9264cb89621a360a7f6e348c63757ac393cc54abff"} Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.455711 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.457239 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.468326 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.507181 4989 generic.go:334] "Generic (PLEG): container finished" podID="497314b7-9c0a-4733-8912-4237e1488b72" containerID="1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e" exitCode=0 Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.507225 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerDied","Data":"1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e"} Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.510228 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.611594 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.611642 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c22m8\" (UniqueName: \"kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.611685 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.713428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.713473 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c22m8\" (UniqueName: \"kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.713497 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.713926 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.714386 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.736595 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c22m8\" (UniqueName: \"kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8\") pod \"certified-operators-4gzsk\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:08 crc kubenswrapper[4989]: I1006 09:10:08.780491 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:09 crc kubenswrapper[4989]: I1006 09:10:09.045051 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:09 crc kubenswrapper[4989]: W1006 09:10:09.062389 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2034adb2_1ab0_4b2a_a98e_e86e6c06cb89.slice/crio-7fd9eed1857ca57664ae932afdb5fb71a7978311e20def2bfea4fda7839dfc39 WatchSource:0}: Error finding container 7fd9eed1857ca57664ae932afdb5fb71a7978311e20def2bfea4fda7839dfc39: Status 404 returned error can't find the container with id 7fd9eed1857ca57664ae932afdb5fb71a7978311e20def2bfea4fda7839dfc39 Oct 06 09:10:09 crc kubenswrapper[4989]: I1006 09:10:09.518307 4989 generic.go:334] "Generic (PLEG): container finished" podID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerID="a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0" exitCode=0 Oct 06 09:10:09 crc kubenswrapper[4989]: I1006 09:10:09.518369 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerDied","Data":"a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0"} Oct 06 09:10:09 crc kubenswrapper[4989]: I1006 09:10:09.518405 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerStarted","Data":"7fd9eed1857ca57664ae932afdb5fb71a7978311e20def2bfea4fda7839dfc39"} Oct 06 09:10:10 crc kubenswrapper[4989]: I1006 09:10:10.529192 4989 generic.go:334] "Generic (PLEG): container finished" podID="497314b7-9c0a-4733-8912-4237e1488b72" containerID="72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9" exitCode=0 Oct 06 09:10:10 crc kubenswrapper[4989]: I1006 09:10:10.529261 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerDied","Data":"72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9"} Oct 06 09:10:10 crc kubenswrapper[4989]: I1006 09:10:10.534521 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerStarted","Data":"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b"} Oct 06 09:10:11 crc kubenswrapper[4989]: I1006 09:10:11.545571 4989 generic.go:334] "Generic (PLEG): container finished" podID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerID="e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b" exitCode=0 Oct 06 09:10:11 crc kubenswrapper[4989]: I1006 09:10:11.545678 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerDied","Data":"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b"} Oct 06 09:10:11 crc kubenswrapper[4989]: I1006 09:10:11.547968 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerStarted","Data":"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5"} Oct 06 09:10:11 crc kubenswrapper[4989]: I1006 09:10:11.595258 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p45wp" podStartSLOduration=3.081101013 podStartE2EDuration="5.595239054s" podCreationTimestamp="2025-10-06 09:10:06 +0000 UTC" firstStartedPulling="2025-10-06 09:10:08.50981585 +0000 UTC m=+1859.299841450" lastFinishedPulling="2025-10-06 09:10:11.023953911 +0000 UTC m=+1861.813979491" observedRunningTime="2025-10-06 09:10:11.593184583 +0000 UTC m=+1862.383210173" watchObservedRunningTime="2025-10-06 09:10:11.595239054 +0000 UTC m=+1862.385264634" Oct 06 09:10:12 crc kubenswrapper[4989]: I1006 09:10:12.558127 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerStarted","Data":"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977"} Oct 06 09:10:12 crc kubenswrapper[4989]: I1006 09:10:12.576199 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4gzsk" podStartSLOduration=2.12683409 podStartE2EDuration="4.576185347s" podCreationTimestamp="2025-10-06 09:10:08 +0000 UTC" firstStartedPulling="2025-10-06 09:10:09.520606781 +0000 UTC m=+1860.310632401" lastFinishedPulling="2025-10-06 09:10:11.969958078 +0000 UTC m=+1862.759983658" observedRunningTime="2025-10-06 09:10:12.572989163 +0000 UTC m=+1863.363014743" watchObservedRunningTime="2025-10-06 09:10:12.576185347 +0000 UTC m=+1863.366210927" Oct 06 09:10:15 crc kubenswrapper[4989]: I1006 09:10:15.936511 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:10:15 crc kubenswrapper[4989]: E1006 09:10:15.936856 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:10:17 crc kubenswrapper[4989]: I1006 09:10:17.007190 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:17 crc kubenswrapper[4989]: I1006 09:10:17.010027 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:17 crc kubenswrapper[4989]: I1006 09:10:17.060312 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:17 crc kubenswrapper[4989]: I1006 09:10:17.661795 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:17 crc kubenswrapper[4989]: I1006 09:10:17.730413 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.186056 4989 scope.go:117] "RemoveContainer" containerID="648b0a633bb695eac1b4b106abd3e4423a4c45eceef2c1366e8d7b5a4a2862d4" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.214627 4989 scope.go:117] "RemoveContainer" containerID="9be8370edf964b02bc2eec83c55e0ec192884f6e6390d307a5c436a87e7f3ff2" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.240538 4989 scope.go:117] "RemoveContainer" containerID="41c3a3ea88eda96a1475512601b4880062d4b09816817a60f531e6bc6a058912" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.283931 4989 scope.go:117] "RemoveContainer" containerID="fa81e6c26183ea833e5e2d25e80e937b2130d7c82f3b53456b03c4b55db2e255" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.310752 4989 scope.go:117] "RemoveContainer" containerID="051491a73335d385cb26783d3784e40625daa8ecb635cd51e999f4f01ba63774" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.336490 4989 scope.go:117] "RemoveContainer" containerID="385a5a4463655d9562fa75c66db444ad5a9ac4c63973728715660d99561581d1" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.385520 4989 scope.go:117] "RemoveContainer" containerID="de97aa68267d73cae2bc02e80acd84df5900e6e32436971bd32cc722e73765e7" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.418779 4989 scope.go:117] "RemoveContainer" containerID="9e2d55d5310b2a032c3053d2f611de15d5d3b47bc88f9275f2f82580ad17ce65" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.442571 4989 scope.go:117] "RemoveContainer" containerID="57bccaac485a945bfa34ccfa394125f43a85083ec91573de318dfac2c1dbab2e" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.781260 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.781324 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:18 crc kubenswrapper[4989]: I1006 09:10:18.829446 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:19 crc kubenswrapper[4989]: I1006 09:10:19.631180 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p45wp" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="registry-server" containerID="cri-o://7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5" gracePeriod=2 Oct 06 09:10:19 crc kubenswrapper[4989]: I1006 09:10:19.718708 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:19 crc kubenswrapper[4989]: I1006 09:10:19.991536 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.081369 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content\") pod \"497314b7-9c0a-4733-8912-4237e1488b72\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.081479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg\") pod \"497314b7-9c0a-4733-8912-4237e1488b72\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.081614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities\") pod \"497314b7-9c0a-4733-8912-4237e1488b72\" (UID: \"497314b7-9c0a-4733-8912-4237e1488b72\") " Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.083247 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities" (OuterVolumeSpecName: "utilities") pod "497314b7-9c0a-4733-8912-4237e1488b72" (UID: "497314b7-9c0a-4733-8912-4237e1488b72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.093593 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg" (OuterVolumeSpecName: "kube-api-access-pjcsg") pod "497314b7-9c0a-4733-8912-4237e1488b72" (UID: "497314b7-9c0a-4733-8912-4237e1488b72"). InnerVolumeSpecName "kube-api-access-pjcsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.142707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "497314b7-9c0a-4733-8912-4237e1488b72" (UID: "497314b7-9c0a-4733-8912-4237e1488b72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.183361 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/497314b7-9c0a-4733-8912-4237e1488b72-kube-api-access-pjcsg\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.183398 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.183407 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/497314b7-9c0a-4733-8912-4237e1488b72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.640435 4989 generic.go:334] "Generic (PLEG): container finished" podID="497314b7-9c0a-4733-8912-4237e1488b72" containerID="7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5" exitCode=0 Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.640525 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p45wp" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.640517 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerDied","Data":"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5"} Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.640605 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p45wp" event={"ID":"497314b7-9c0a-4733-8912-4237e1488b72","Type":"ContainerDied","Data":"44a0a3b69b20ffbea65bfd9264cb89621a360a7f6e348c63757ac393cc54abff"} Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.640636 4989 scope.go:117] "RemoveContainer" containerID="7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.664505 4989 scope.go:117] "RemoveContainer" containerID="72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.685290 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.689629 4989 scope.go:117] "RemoveContainer" containerID="1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.690022 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p45wp"] Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.727891 4989 scope.go:117] "RemoveContainer" containerID="7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5" Oct 06 09:10:20 crc kubenswrapper[4989]: E1006 09:10:20.728280 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5\": container with ID starting with 7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5 not found: ID does not exist" containerID="7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.728312 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5"} err="failed to get container status \"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5\": rpc error: code = NotFound desc = could not find container \"7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5\": container with ID starting with 7d64b3062512e4df7d3ab2ce4b10a3776dd1a1d16595ed72ed6befae89f6aec5 not found: ID does not exist" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.728643 4989 scope.go:117] "RemoveContainer" containerID="72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9" Oct 06 09:10:20 crc kubenswrapper[4989]: E1006 09:10:20.728885 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9\": container with ID starting with 72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9 not found: ID does not exist" containerID="72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.728921 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9"} err="failed to get container status \"72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9\": rpc error: code = NotFound desc = could not find container \"72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9\": container with ID starting with 72a84f7d9d46878e97bbeac7f7b869edf628785521a5c448e313a2004149d1d9 not found: ID does not exist" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.728941 4989 scope.go:117] "RemoveContainer" containerID="1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e" Oct 06 09:10:20 crc kubenswrapper[4989]: E1006 09:10:20.729124 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e\": container with ID starting with 1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e not found: ID does not exist" containerID="1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e" Oct 06 09:10:20 crc kubenswrapper[4989]: I1006 09:10:20.729160 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e"} err="failed to get container status \"1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e\": rpc error: code = NotFound desc = could not find container \"1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e\": container with ID starting with 1175b03d87cc49a6884b7d2e5b3511e1a3b3008de71e40a18c6e3618ee61897e not found: ID does not exist" Oct 06 09:10:21 crc kubenswrapper[4989]: I1006 09:10:21.297752 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:21 crc kubenswrapper[4989]: I1006 09:10:21.652081 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4gzsk" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="registry-server" containerID="cri-o://3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977" gracePeriod=2 Oct 06 09:10:21 crc kubenswrapper[4989]: I1006 09:10:21.953827 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="497314b7-9c0a-4733-8912-4237e1488b72" path="/var/lib/kubelet/pods/497314b7-9c0a-4733-8912-4237e1488b72/volumes" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.035742 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.211471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c22m8\" (UniqueName: \"kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8\") pod \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.211587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities\") pod \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.211759 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content\") pod \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\" (UID: \"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89\") " Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.213385 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities" (OuterVolumeSpecName: "utilities") pod "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" (UID: "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.217554 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8" (OuterVolumeSpecName: "kube-api-access-c22m8") pod "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" (UID: "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89"). InnerVolumeSpecName "kube-api-access-c22m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.268544 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" (UID: "2034adb2-1ab0-4b2a-a98e-e86e6c06cb89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.313670 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.313703 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.313724 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c22m8\" (UniqueName: \"kubernetes.io/projected/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89-kube-api-access-c22m8\") on node \"crc\" DevicePath \"\"" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.678550 4989 generic.go:334] "Generic (PLEG): container finished" podID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerID="3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977" exitCode=0 Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.678646 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4gzsk" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.678686 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerDied","Data":"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977"} Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.678767 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4gzsk" event={"ID":"2034adb2-1ab0-4b2a-a98e-e86e6c06cb89","Type":"ContainerDied","Data":"7fd9eed1857ca57664ae932afdb5fb71a7978311e20def2bfea4fda7839dfc39"} Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.678798 4989 scope.go:117] "RemoveContainer" containerID="3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.722893 4989 scope.go:117] "RemoveContainer" containerID="e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.728089 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.736895 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4gzsk"] Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.750516 4989 scope.go:117] "RemoveContainer" containerID="a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.785059 4989 scope.go:117] "RemoveContainer" containerID="3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977" Oct 06 09:10:22 crc kubenswrapper[4989]: E1006 09:10:22.785506 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977\": container with ID starting with 3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977 not found: ID does not exist" containerID="3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.785563 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977"} err="failed to get container status \"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977\": rpc error: code = NotFound desc = could not find container \"3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977\": container with ID starting with 3dc8331bfcd3066d593cb5ae0e6a5bad1cb348369dc18f1a5dd51af4681d9977 not found: ID does not exist" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.785604 4989 scope.go:117] "RemoveContainer" containerID="e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b" Oct 06 09:10:22 crc kubenswrapper[4989]: E1006 09:10:22.785957 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b\": container with ID starting with e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b not found: ID does not exist" containerID="e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.785998 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b"} err="failed to get container status \"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b\": rpc error: code = NotFound desc = could not find container \"e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b\": container with ID starting with e9b457590a813e719c662e8efc3f6492a4f576da208602b1da89ed538113119b not found: ID does not exist" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.786017 4989 scope.go:117] "RemoveContainer" containerID="a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0" Oct 06 09:10:22 crc kubenswrapper[4989]: E1006 09:10:22.786310 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0\": container with ID starting with a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0 not found: ID does not exist" containerID="a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0" Oct 06 09:10:22 crc kubenswrapper[4989]: I1006 09:10:22.786338 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0"} err="failed to get container status \"a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0\": rpc error: code = NotFound desc = could not find container \"a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0\": container with ID starting with a8e49f3fba72e3a24f02df1065f8e743b4f2c7c26279da7ac8ba987530ded9a0 not found: ID does not exist" Oct 06 09:10:23 crc kubenswrapper[4989]: I1006 09:10:23.945218 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" path="/var/lib/kubelet/pods/2034adb2-1ab0-4b2a-a98e-e86e6c06cb89/volumes" Oct 06 09:10:30 crc kubenswrapper[4989]: I1006 09:10:30.937540 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:10:30 crc kubenswrapper[4989]: E1006 09:10:30.938355 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:10:43 crc kubenswrapper[4989]: I1006 09:10:43.936098 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:10:43 crc kubenswrapper[4989]: E1006 09:10:43.936784 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:10:58 crc kubenswrapper[4989]: I1006 09:10:58.936286 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:10:58 crc kubenswrapper[4989]: E1006 09:10:58.938349 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:11:13 crc kubenswrapper[4989]: I1006 09:11:13.938064 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:11:13 crc kubenswrapper[4989]: E1006 09:11:13.939377 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:11:25 crc kubenswrapper[4989]: I1006 09:11:25.936205 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:11:25 crc kubenswrapper[4989]: E1006 09:11:25.938318 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:11:39 crc kubenswrapper[4989]: I1006 09:11:39.940599 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:11:39 crc kubenswrapper[4989]: E1006 09:11:39.941443 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:11:54 crc kubenswrapper[4989]: I1006 09:11:54.936642 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:11:54 crc kubenswrapper[4989]: E1006 09:11:54.937462 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:12:06 crc kubenswrapper[4989]: I1006 09:12:06.935592 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:12:06 crc kubenswrapper[4989]: E1006 09:12:06.936535 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:12:17 crc kubenswrapper[4989]: I1006 09:12:17.936130 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:12:17 crc kubenswrapper[4989]: E1006 09:12:17.937252 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:12:28 crc kubenswrapper[4989]: I1006 09:12:28.936372 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:12:28 crc kubenswrapper[4989]: E1006 09:12:28.937197 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:12:40 crc kubenswrapper[4989]: I1006 09:12:40.936216 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:12:41 crc kubenswrapper[4989]: I1006 09:12:41.810211 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b"} Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.420561 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424279 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424316 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424332 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="extract-utilities" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424340 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="extract-utilities" Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424351 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="extract-content" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424358 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="extract-content" Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424370 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="extract-utilities" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424377 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="extract-utilities" Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424386 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="extract-content" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424395 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="extract-content" Oct 06 09:13:33 crc kubenswrapper[4989]: E1006 09:13:33.424411 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424418 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424607 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="497314b7-9c0a-4733-8912-4237e1488b72" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.424619 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2034adb2-1ab0-4b2a-a98e-e86e6c06cb89" containerName="registry-server" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.425924 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.460947 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.595777 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.595871 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g24zh\" (UniqueName: \"kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.595925 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.697204 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.697265 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g24zh\" (UniqueName: \"kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.697321 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.698012 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.698007 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.719944 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g24zh\" (UniqueName: \"kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh\") pod \"redhat-marketplace-7jc2z\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:33 crc kubenswrapper[4989]: I1006 09:13:33.763934 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:34 crc kubenswrapper[4989]: I1006 09:13:34.175625 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:34 crc kubenswrapper[4989]: I1006 09:13:34.222227 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerStarted","Data":"86f7ee0d17d8e6215eee55e28ea5773d40e9bf4c6496ad64c2dc7261c52ff530"} Oct 06 09:13:35 crc kubenswrapper[4989]: I1006 09:13:35.231510 4989 generic.go:334] "Generic (PLEG): container finished" podID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerID="961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564" exitCode=0 Oct 06 09:13:35 crc kubenswrapper[4989]: I1006 09:13:35.231584 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerDied","Data":"961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564"} Oct 06 09:13:36 crc kubenswrapper[4989]: I1006 09:13:36.240777 4989 generic.go:334] "Generic (PLEG): container finished" podID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerID="711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d" exitCode=0 Oct 06 09:13:36 crc kubenswrapper[4989]: I1006 09:13:36.240818 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerDied","Data":"711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d"} Oct 06 09:13:37 crc kubenswrapper[4989]: I1006 09:13:37.249210 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerStarted","Data":"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c"} Oct 06 09:13:43 crc kubenswrapper[4989]: I1006 09:13:43.764560 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:43 crc kubenswrapper[4989]: I1006 09:13:43.765859 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:43 crc kubenswrapper[4989]: I1006 09:13:43.831414 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:43 crc kubenswrapper[4989]: I1006 09:13:43.857952 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7jc2z" podStartSLOduration=9.404276201 podStartE2EDuration="10.85792858s" podCreationTimestamp="2025-10-06 09:13:33 +0000 UTC" firstStartedPulling="2025-10-06 09:13:35.233375954 +0000 UTC m=+2066.023401534" lastFinishedPulling="2025-10-06 09:13:36.687028333 +0000 UTC m=+2067.477053913" observedRunningTime="2025-10-06 09:13:37.273864154 +0000 UTC m=+2068.063889734" watchObservedRunningTime="2025-10-06 09:13:43.85792858 +0000 UTC m=+2074.647954190" Oct 06 09:13:44 crc kubenswrapper[4989]: I1006 09:13:44.361942 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:44 crc kubenswrapper[4989]: I1006 09:13:44.432169 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.310920 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7jc2z" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="registry-server" containerID="cri-o://3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c" gracePeriod=2 Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.729114 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.891837 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities\") pod \"ebd6005e-4e76-4a91-a57b-d74b6703a344\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.892132 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g24zh\" (UniqueName: \"kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh\") pod \"ebd6005e-4e76-4a91-a57b-d74b6703a344\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.892242 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content\") pod \"ebd6005e-4e76-4a91-a57b-d74b6703a344\" (UID: \"ebd6005e-4e76-4a91-a57b-d74b6703a344\") " Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.893574 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities" (OuterVolumeSpecName: "utilities") pod "ebd6005e-4e76-4a91-a57b-d74b6703a344" (UID: "ebd6005e-4e76-4a91-a57b-d74b6703a344"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.898459 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh" (OuterVolumeSpecName: "kube-api-access-g24zh") pod "ebd6005e-4e76-4a91-a57b-d74b6703a344" (UID: "ebd6005e-4e76-4a91-a57b-d74b6703a344"). InnerVolumeSpecName "kube-api-access-g24zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.906835 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebd6005e-4e76-4a91-a57b-d74b6703a344" (UID: "ebd6005e-4e76-4a91-a57b-d74b6703a344"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.994787 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.994828 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebd6005e-4e76-4a91-a57b-d74b6703a344-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:13:46 crc kubenswrapper[4989]: I1006 09:13:46.994843 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g24zh\" (UniqueName: \"kubernetes.io/projected/ebd6005e-4e76-4a91-a57b-d74b6703a344-kube-api-access-g24zh\") on node \"crc\" DevicePath \"\"" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.328469 4989 generic.go:334] "Generic (PLEG): container finished" podID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerID="3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c" exitCode=0 Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.328511 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerDied","Data":"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c"} Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.328536 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7jc2z" event={"ID":"ebd6005e-4e76-4a91-a57b-d74b6703a344","Type":"ContainerDied","Data":"86f7ee0d17d8e6215eee55e28ea5773d40e9bf4c6496ad64c2dc7261c52ff530"} Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.328552 4989 scope.go:117] "RemoveContainer" containerID="3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.328688 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7jc2z" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.345055 4989 scope.go:117] "RemoveContainer" containerID="711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.358402 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.364877 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7jc2z"] Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.366342 4989 scope.go:117] "RemoveContainer" containerID="961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.393626 4989 scope.go:117] "RemoveContainer" containerID="3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c" Oct 06 09:13:47 crc kubenswrapper[4989]: E1006 09:13:47.394078 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c\": container with ID starting with 3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c not found: ID does not exist" containerID="3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.394115 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c"} err="failed to get container status \"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c\": rpc error: code = NotFound desc = could not find container \"3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c\": container with ID starting with 3351ccb9a44360816bf891cd7796ce6efd6e116657f28a9ff9492720c44c6e8c not found: ID does not exist" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.394140 4989 scope.go:117] "RemoveContainer" containerID="711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d" Oct 06 09:13:47 crc kubenswrapper[4989]: E1006 09:13:47.394421 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d\": container with ID starting with 711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d not found: ID does not exist" containerID="711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.394473 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d"} err="failed to get container status \"711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d\": rpc error: code = NotFound desc = could not find container \"711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d\": container with ID starting with 711892052ecd8dbdd1d16b18d33560e42136e21ab21e3964d12dd6d00008fe0d not found: ID does not exist" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.394505 4989 scope.go:117] "RemoveContainer" containerID="961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564" Oct 06 09:13:47 crc kubenswrapper[4989]: E1006 09:13:47.394807 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564\": container with ID starting with 961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564 not found: ID does not exist" containerID="961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.394841 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564"} err="failed to get container status \"961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564\": rpc error: code = NotFound desc = could not find container \"961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564\": container with ID starting with 961259125baaec24b8f0a43058fd4159d736835cef787fe0cc702135f9e28564 not found: ID does not exist" Oct 06 09:13:47 crc kubenswrapper[4989]: I1006 09:13:47.948790 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" path="/var/lib/kubelet/pods/ebd6005e-4e76-4a91-a57b-d74b6703a344/volumes" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.140335 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k"] Oct 06 09:15:00 crc kubenswrapper[4989]: E1006 09:15:00.141217 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="registry-server" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.141235 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="registry-server" Oct 06 09:15:00 crc kubenswrapper[4989]: E1006 09:15:00.141261 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="extract-content" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.141269 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="extract-content" Oct 06 09:15:00 crc kubenswrapper[4989]: E1006 09:15:00.141282 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="extract-utilities" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.141290 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="extract-utilities" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.141468 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd6005e-4e76-4a91-a57b-d74b6703a344" containerName="registry-server" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.142079 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.144417 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.144718 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.157487 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k"] Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.261207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.261308 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxlzw\" (UniqueName: \"kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.261358 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.362781 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.362873 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxlzw\" (UniqueName: \"kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.362984 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.364272 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.370851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.386926 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxlzw\" (UniqueName: \"kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw\") pod \"collect-profiles-29329035-lsz9k\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.480758 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:00 crc kubenswrapper[4989]: I1006 09:15:00.922619 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k"] Oct 06 09:15:01 crc kubenswrapper[4989]: I1006 09:15:01.896016 4989 generic.go:334] "Generic (PLEG): container finished" podID="b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" containerID="b5382f952c8fad7bb2d79085bc023bbfaeb873c6af55a666a343e264c10d3f47" exitCode=0 Oct 06 09:15:01 crc kubenswrapper[4989]: I1006 09:15:01.896060 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" event={"ID":"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72","Type":"ContainerDied","Data":"b5382f952c8fad7bb2d79085bc023bbfaeb873c6af55a666a343e264c10d3f47"} Oct 06 09:15:01 crc kubenswrapper[4989]: I1006 09:15:01.896312 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" event={"ID":"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72","Type":"ContainerStarted","Data":"1ed8992dcd4a902d51a4ef3d24ad297aa39a2d6a7fe594f03562239b580ed804"} Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.185097 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.304845 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume\") pod \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.305002 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxlzw\" (UniqueName: \"kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw\") pod \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.305090 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume\") pod \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\" (UID: \"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72\") " Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.305646 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume" (OuterVolumeSpecName: "config-volume") pod "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" (UID: "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.310857 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw" (OuterVolumeSpecName: "kube-api-access-rxlzw") pod "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" (UID: "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72"). InnerVolumeSpecName "kube-api-access-rxlzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.310968 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" (UID: "b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.406893 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxlzw\" (UniqueName: \"kubernetes.io/projected/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-kube-api-access-rxlzw\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.406940 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.406952 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.909286 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" event={"ID":"b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72","Type":"ContainerDied","Data":"1ed8992dcd4a902d51a4ef3d24ad297aa39a2d6a7fe594f03562239b580ed804"} Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.909329 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed8992dcd4a902d51a4ef3d24ad297aa39a2d6a7fe594f03562239b580ed804" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.909341 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k" Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.939202 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:15:03 crc kubenswrapper[4989]: I1006 09:15:03.939264 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:15:04 crc kubenswrapper[4989]: I1006 09:15:04.267449 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88"] Oct 06 09:15:04 crc kubenswrapper[4989]: I1006 09:15:04.275708 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29328990-v4h88"] Oct 06 09:15:05 crc kubenswrapper[4989]: I1006 09:15:05.947489 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ae2464-66b2-4986-beef-382caadb447e" path="/var/lib/kubelet/pods/25ae2464-66b2-4986-beef-382caadb447e/volumes" Oct 06 09:15:18 crc kubenswrapper[4989]: I1006 09:15:18.671127 4989 scope.go:117] "RemoveContainer" containerID="4d953e5712bd3c81aab1f7118d30d60f3d549a967fbebf044b5f14a3ce7546c3" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.351435 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:32 crc kubenswrapper[4989]: E1006 09:15:32.352776 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" containerName="collect-profiles" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.352808 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" containerName="collect-profiles" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.353185 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" containerName="collect-profiles" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.358996 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.370326 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.439628 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.439701 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.439887 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5g2d\" (UniqueName: \"kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.541499 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.542087 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.542014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.542204 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5g2d\" (UniqueName: \"kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.542484 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.573476 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5g2d\" (UniqueName: \"kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d\") pod \"redhat-operators-z4f7k\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:32 crc kubenswrapper[4989]: I1006 09:15:32.689527 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:33 crc kubenswrapper[4989]: I1006 09:15:33.099689 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:33 crc kubenswrapper[4989]: I1006 09:15:33.139008 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerStarted","Data":"b64d9faa298282d1e5f1647007c6cdac94d7fa9bc25ad348e91fdde8702914d0"} Oct 06 09:15:33 crc kubenswrapper[4989]: I1006 09:15:33.936068 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:15:33 crc kubenswrapper[4989]: I1006 09:15:33.936397 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:15:34 crc kubenswrapper[4989]: I1006 09:15:34.148564 4989 generic.go:334] "Generic (PLEG): container finished" podID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerID="a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d" exitCode=0 Oct 06 09:15:34 crc kubenswrapper[4989]: I1006 09:15:34.148627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerDied","Data":"a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d"} Oct 06 09:15:34 crc kubenswrapper[4989]: I1006 09:15:34.151051 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:15:36 crc kubenswrapper[4989]: I1006 09:15:36.168318 4989 generic.go:334] "Generic (PLEG): container finished" podID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerID="303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0" exitCode=0 Oct 06 09:15:36 crc kubenswrapper[4989]: I1006 09:15:36.168477 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerDied","Data":"303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0"} Oct 06 09:15:37 crc kubenswrapper[4989]: I1006 09:15:37.181334 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerStarted","Data":"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3"} Oct 06 09:15:37 crc kubenswrapper[4989]: I1006 09:15:37.203348 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z4f7k" podStartSLOduration=2.587471539 podStartE2EDuration="5.203322567s" podCreationTimestamp="2025-10-06 09:15:32 +0000 UTC" firstStartedPulling="2025-10-06 09:15:34.150796325 +0000 UTC m=+2184.940821905" lastFinishedPulling="2025-10-06 09:15:36.766647353 +0000 UTC m=+2187.556672933" observedRunningTime="2025-10-06 09:15:37.196799577 +0000 UTC m=+2187.986825177" watchObservedRunningTime="2025-10-06 09:15:37.203322567 +0000 UTC m=+2187.993348187" Oct 06 09:15:42 crc kubenswrapper[4989]: I1006 09:15:42.690246 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:42 crc kubenswrapper[4989]: I1006 09:15:42.690354 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:42 crc kubenswrapper[4989]: I1006 09:15:42.739516 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:43 crc kubenswrapper[4989]: I1006 09:15:43.292219 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:43 crc kubenswrapper[4989]: I1006 09:15:43.358777 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.250752 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z4f7k" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="registry-server" containerID="cri-o://ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3" gracePeriod=2 Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.686773 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.737009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities\") pod \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.737107 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content\") pod \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.737150 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5g2d\" (UniqueName: \"kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d\") pod \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\" (UID: \"76d8a919-ba3c-4e18-bebb-bf623a1f39a1\") " Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.737962 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities" (OuterVolumeSpecName: "utilities") pod "76d8a919-ba3c-4e18-bebb-bf623a1f39a1" (UID: "76d8a919-ba3c-4e18-bebb-bf623a1f39a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.742781 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d" (OuterVolumeSpecName: "kube-api-access-x5g2d") pod "76d8a919-ba3c-4e18-bebb-bf623a1f39a1" (UID: "76d8a919-ba3c-4e18-bebb-bf623a1f39a1"). InnerVolumeSpecName "kube-api-access-x5g2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.839258 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5g2d\" (UniqueName: \"kubernetes.io/projected/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-kube-api-access-x5g2d\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.839300 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.842617 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76d8a919-ba3c-4e18-bebb-bf623a1f39a1" (UID: "76d8a919-ba3c-4e18-bebb-bf623a1f39a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:15:45 crc kubenswrapper[4989]: I1006 09:15:45.940099 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d8a919-ba3c-4e18-bebb-bf623a1f39a1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.262984 4989 generic.go:334] "Generic (PLEG): container finished" podID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerID="ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3" exitCode=0 Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.263037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerDied","Data":"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3"} Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.263073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z4f7k" event={"ID":"76d8a919-ba3c-4e18-bebb-bf623a1f39a1","Type":"ContainerDied","Data":"b64d9faa298282d1e5f1647007c6cdac94d7fa9bc25ad348e91fdde8702914d0"} Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.263094 4989 scope.go:117] "RemoveContainer" containerID="ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.263239 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z4f7k" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.287495 4989 scope.go:117] "RemoveContainer" containerID="303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.294794 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.300502 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z4f7k"] Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.305149 4989 scope.go:117] "RemoveContainer" containerID="a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.328055 4989 scope.go:117] "RemoveContainer" containerID="ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3" Oct 06 09:15:46 crc kubenswrapper[4989]: E1006 09:15:46.328545 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3\": container with ID starting with ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3 not found: ID does not exist" containerID="ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.328591 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3"} err="failed to get container status \"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3\": rpc error: code = NotFound desc = could not find container \"ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3\": container with ID starting with ff3729d512e1d88d9dc42ba210e222d766ed65347c0840e3304453a4c32c80a3 not found: ID does not exist" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.328618 4989 scope.go:117] "RemoveContainer" containerID="303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0" Oct 06 09:15:46 crc kubenswrapper[4989]: E1006 09:15:46.329036 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0\": container with ID starting with 303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0 not found: ID does not exist" containerID="303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.329091 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0"} err="failed to get container status \"303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0\": rpc error: code = NotFound desc = could not find container \"303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0\": container with ID starting with 303c4ceb46d4963a3d35b722a36db374881ad9744d8c0ca284a17f47f3e19cc0 not found: ID does not exist" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.329118 4989 scope.go:117] "RemoveContainer" containerID="a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d" Oct 06 09:15:46 crc kubenswrapper[4989]: E1006 09:15:46.329388 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d\": container with ID starting with a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d not found: ID does not exist" containerID="a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d" Oct 06 09:15:46 crc kubenswrapper[4989]: I1006 09:15:46.329415 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d"} err="failed to get container status \"a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d\": rpc error: code = NotFound desc = could not find container \"a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d\": container with ID starting with a81dad5da3500671c56d13925c8497d6e9fb683089a5437225599846dc7e275d not found: ID does not exist" Oct 06 09:15:47 crc kubenswrapper[4989]: I1006 09:15:47.948784 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" path="/var/lib/kubelet/pods/76d8a919-ba3c-4e18-bebb-bf623a1f39a1/volumes" Oct 06 09:16:03 crc kubenswrapper[4989]: I1006 09:16:03.935334 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:16:03 crc kubenswrapper[4989]: I1006 09:16:03.936313 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:16:03 crc kubenswrapper[4989]: I1006 09:16:03.946428 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:16:03 crc kubenswrapper[4989]: I1006 09:16:03.947117 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:16:03 crc kubenswrapper[4989]: I1006 09:16:03.947188 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b" gracePeriod=600 Oct 06 09:16:04 crc kubenswrapper[4989]: I1006 09:16:04.410902 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b" exitCode=0 Oct 06 09:16:04 crc kubenswrapper[4989]: I1006 09:16:04.410954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b"} Oct 06 09:16:04 crc kubenswrapper[4989]: I1006 09:16:04.411572 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98"} Oct 06 09:16:04 crc kubenswrapper[4989]: I1006 09:16:04.411629 4989 scope.go:117] "RemoveContainer" containerID="d9a3326054da16e938dfc8117ddc401b3a5f180be246e921d3959d452c2fcaf6" Oct 06 09:18:33 crc kubenswrapper[4989]: I1006 09:18:33.935393 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:18:33 crc kubenswrapper[4989]: I1006 09:18:33.936009 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:19:03 crc kubenswrapper[4989]: I1006 09:19:03.935811 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:19:03 crc kubenswrapper[4989]: I1006 09:19:03.936497 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:19:33 crc kubenswrapper[4989]: I1006 09:19:33.935612 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:19:33 crc kubenswrapper[4989]: I1006 09:19:33.936413 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:19:33 crc kubenswrapper[4989]: I1006 09:19:33.944743 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:19:33 crc kubenswrapper[4989]: I1006 09:19:33.945448 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:19:33 crc kubenswrapper[4989]: I1006 09:19:33.945533 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" gracePeriod=600 Oct 06 09:19:34 crc kubenswrapper[4989]: E1006 09:19:34.067896 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:19:35 crc kubenswrapper[4989]: I1006 09:19:35.044774 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" exitCode=0 Oct 06 09:19:35 crc kubenswrapper[4989]: I1006 09:19:35.044841 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98"} Oct 06 09:19:35 crc kubenswrapper[4989]: I1006 09:19:35.045155 4989 scope.go:117] "RemoveContainer" containerID="e00750ce96bf25691a8ebc8e2744a59478ab68f3146189f81ecd36bf1dd4b62b" Oct 06 09:19:35 crc kubenswrapper[4989]: I1006 09:19:35.045449 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:19:35 crc kubenswrapper[4989]: E1006 09:19:35.045694 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:19:45 crc kubenswrapper[4989]: I1006 09:19:45.935269 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:19:45 crc kubenswrapper[4989]: E1006 09:19:45.937133 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:19:58 crc kubenswrapper[4989]: I1006 09:19:58.937245 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:19:58 crc kubenswrapper[4989]: E1006 09:19:58.938402 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:20:10 crc kubenswrapper[4989]: I1006 09:20:10.935635 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:20:10 crc kubenswrapper[4989]: E1006 09:20:10.936455 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:20:23 crc kubenswrapper[4989]: I1006 09:20:23.937544 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:20:23 crc kubenswrapper[4989]: E1006 09:20:23.938890 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:20:36 crc kubenswrapper[4989]: I1006 09:20:36.936122 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:20:36 crc kubenswrapper[4989]: E1006 09:20:36.936945 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:20:47 crc kubenswrapper[4989]: I1006 09:20:47.936879 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:20:47 crc kubenswrapper[4989]: E1006 09:20:47.937940 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:00 crc kubenswrapper[4989]: I1006 09:21:00.935690 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:00 crc kubenswrapper[4989]: E1006 09:21:00.936894 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:11 crc kubenswrapper[4989]: I1006 09:21:11.936573 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:11 crc kubenswrapper[4989]: E1006 09:21:11.937619 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:22 crc kubenswrapper[4989]: I1006 09:21:22.936057 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:22 crc kubenswrapper[4989]: E1006 09:21:22.936680 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:33 crc kubenswrapper[4989]: I1006 09:21:33.935824 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:33 crc kubenswrapper[4989]: E1006 09:21:33.936543 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:44 crc kubenswrapper[4989]: I1006 09:21:44.935833 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:44 crc kubenswrapper[4989]: E1006 09:21:44.936580 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:21:57 crc kubenswrapper[4989]: I1006 09:21:57.935591 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:21:57 crc kubenswrapper[4989]: E1006 09:21:57.936403 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:22:08 crc kubenswrapper[4989]: I1006 09:22:08.936700 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:22:08 crc kubenswrapper[4989]: E1006 09:22:08.939495 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:22:23 crc kubenswrapper[4989]: I1006 09:22:23.936897 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:22:23 crc kubenswrapper[4989]: E1006 09:22:23.937722 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:22:34 crc kubenswrapper[4989]: I1006 09:22:34.935396 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:22:34 crc kubenswrapper[4989]: E1006 09:22:34.936079 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:22:47 crc kubenswrapper[4989]: I1006 09:22:47.937134 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:22:47 crc kubenswrapper[4989]: E1006 09:22:47.938314 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:23:00 crc kubenswrapper[4989]: I1006 09:23:00.937197 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:23:00 crc kubenswrapper[4989]: E1006 09:23:00.938437 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:23:11 crc kubenswrapper[4989]: I1006 09:23:11.935511 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:23:11 crc kubenswrapper[4989]: E1006 09:23:11.936282 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:23:23 crc kubenswrapper[4989]: I1006 09:23:23.936756 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:23:23 crc kubenswrapper[4989]: E1006 09:23:23.937679 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:23:35 crc kubenswrapper[4989]: I1006 09:23:35.936848 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:23:35 crc kubenswrapper[4989]: E1006 09:23:35.938588 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:23:48 crc kubenswrapper[4989]: I1006 09:23:48.936425 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:23:48 crc kubenswrapper[4989]: E1006 09:23:48.937577 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:24:02 crc kubenswrapper[4989]: I1006 09:24:02.935448 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:24:02 crc kubenswrapper[4989]: E1006 09:24:02.937054 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:24:13 crc kubenswrapper[4989]: I1006 09:24:13.936007 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:24:13 crc kubenswrapper[4989]: E1006 09:24:13.937117 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.037785 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:17 crc kubenswrapper[4989]: E1006 09:24:17.038818 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="extract-content" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.038847 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="extract-content" Oct 06 09:24:17 crc kubenswrapper[4989]: E1006 09:24:17.038891 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="extract-utilities" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.038908 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="extract-utilities" Oct 06 09:24:17 crc kubenswrapper[4989]: E1006 09:24:17.038938 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="registry-server" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.038956 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="registry-server" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.039347 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d8a919-ba3c-4e18-bebb-bf623a1f39a1" containerName="registry-server" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.041807 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.057438 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.204851 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6p2r\" (UniqueName: \"kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.204924 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.205103 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.221747 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.223348 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.236353 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306297 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306348 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306377 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6p2r\" (UniqueName: \"kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306498 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qvdz\" (UniqueName: \"kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.306860 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.308844 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.328193 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6p2r\" (UniqueName: \"kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r\") pod \"redhat-marketplace-stv6t\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.372821 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.407291 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.407478 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.407635 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qvdz\" (UniqueName: \"kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.407932 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.408065 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.441640 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qvdz\" (UniqueName: \"kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz\") pod \"community-operators-t5j6x\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.550764 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.665871 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:17 crc kubenswrapper[4989]: I1006 09:24:17.900643 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:17 crc kubenswrapper[4989]: W1006 09:24:17.975746 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode78cfbdd_cad3_46e5_9c56_e64e2595adfa.slice/crio-96bc92cdeb000a1799304aae4a2946a45f5946f6681bcc00c2ac90cec302c3a6 WatchSource:0}: Error finding container 96bc92cdeb000a1799304aae4a2946a45f5946f6681bcc00c2ac90cec302c3a6: Status 404 returned error can't find the container with id 96bc92cdeb000a1799304aae4a2946a45f5946f6681bcc00c2ac90cec302c3a6 Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.405295 4989 generic.go:334] "Generic (PLEG): container finished" podID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerID="23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08" exitCode=0 Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.405374 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerDied","Data":"23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08"} Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.405400 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerStarted","Data":"96bc92cdeb000a1799304aae4a2946a45f5946f6681bcc00c2ac90cec302c3a6"} Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.407616 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.408153 4989 generic.go:334] "Generic (PLEG): container finished" podID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerID="de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904" exitCode=0 Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.408177 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerDied","Data":"de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904"} Oct 06 09:24:18 crc kubenswrapper[4989]: I1006 09:24:18.408195 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerStarted","Data":"508d5d146e1104577a87d4a54d60d52618641b9f421afca866447fcf0c13353d"} Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.429037 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.431642 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.435784 4989 generic.go:334] "Generic (PLEG): container finished" podID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerID="06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5" exitCode=0 Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.435839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerDied","Data":"06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5"} Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.437298 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.446945 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx2wm\" (UniqueName: \"kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.446982 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.447040 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.547979 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx2wm\" (UniqueName: \"kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.548032 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.548447 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.548510 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.548951 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.569583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx2wm\" (UniqueName: \"kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm\") pod \"certified-operators-6mgp7\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:19 crc kubenswrapper[4989]: I1006 09:24:19.772396 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.039503 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:20 crc kubenswrapper[4989]: W1006 09:24:20.045046 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e8e26d6_e46a_4c4d_bf50_45da05934bbc.slice/crio-da53cbb078319736db75c04588d11df00dc40a9022d759e046d58a1f84808797 WatchSource:0}: Error finding container da53cbb078319736db75c04588d11df00dc40a9022d759e046d58a1f84808797: Status 404 returned error can't find the container with id da53cbb078319736db75c04588d11df00dc40a9022d759e046d58a1f84808797 Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.445029 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerStarted","Data":"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b"} Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.447532 4989 generic.go:334] "Generic (PLEG): container finished" podID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerID="f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd" exitCode=0 Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.447567 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerDied","Data":"f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd"} Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.449850 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerID="66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6" exitCode=0 Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.449891 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerDied","Data":"66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6"} Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.449913 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerStarted","Data":"da53cbb078319736db75c04588d11df00dc40a9022d759e046d58a1f84808797"} Oct 06 09:24:20 crc kubenswrapper[4989]: I1006 09:24:20.470767 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-stv6t" podStartSLOduration=1.9899350180000002 podStartE2EDuration="3.470744359s" podCreationTimestamp="2025-10-06 09:24:17 +0000 UTC" firstStartedPulling="2025-10-06 09:24:18.413968202 +0000 UTC m=+2709.203993802" lastFinishedPulling="2025-10-06 09:24:19.894777553 +0000 UTC m=+2710.684803143" observedRunningTime="2025-10-06 09:24:20.467525636 +0000 UTC m=+2711.257551236" watchObservedRunningTime="2025-10-06 09:24:20.470744359 +0000 UTC m=+2711.260769939" Oct 06 09:24:21 crc kubenswrapper[4989]: I1006 09:24:21.462510 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerStarted","Data":"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806"} Oct 06 09:24:21 crc kubenswrapper[4989]: I1006 09:24:21.482706 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t5j6x" podStartSLOduration=2.020430598 podStartE2EDuration="4.482642744s" podCreationTimestamp="2025-10-06 09:24:17 +0000 UTC" firstStartedPulling="2025-10-06 09:24:18.40731572 +0000 UTC m=+2709.197341300" lastFinishedPulling="2025-10-06 09:24:20.869527866 +0000 UTC m=+2711.659553446" observedRunningTime="2025-10-06 09:24:21.481308396 +0000 UTC m=+2712.271334016" watchObservedRunningTime="2025-10-06 09:24:21.482642744 +0000 UTC m=+2712.272668324" Oct 06 09:24:22 crc kubenswrapper[4989]: I1006 09:24:22.470459 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerID="8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205" exitCode=0 Oct 06 09:24:22 crc kubenswrapper[4989]: I1006 09:24:22.470570 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerDied","Data":"8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205"} Oct 06 09:24:23 crc kubenswrapper[4989]: I1006 09:24:23.481635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerStarted","Data":"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947"} Oct 06 09:24:23 crc kubenswrapper[4989]: I1006 09:24:23.508631 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6mgp7" podStartSLOduration=1.8894631629999998 podStartE2EDuration="4.508612791s" podCreationTimestamp="2025-10-06 09:24:19 +0000 UTC" firstStartedPulling="2025-10-06 09:24:20.451371069 +0000 UTC m=+2711.241396649" lastFinishedPulling="2025-10-06 09:24:23.070520697 +0000 UTC m=+2713.860546277" observedRunningTime="2025-10-06 09:24:23.501953659 +0000 UTC m=+2714.291979249" watchObservedRunningTime="2025-10-06 09:24:23.508612791 +0000 UTC m=+2714.298638381" Oct 06 09:24:24 crc kubenswrapper[4989]: I1006 09:24:24.936085 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:24:24 crc kubenswrapper[4989]: E1006 09:24:24.936450 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.373414 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.373853 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.430277 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.552024 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.552081 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.565550 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:27 crc kubenswrapper[4989]: I1006 09:24:27.602938 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:28 crc kubenswrapper[4989]: I1006 09:24:28.575192 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.411782 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.533182 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-stv6t" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="registry-server" containerID="cri-o://a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b" gracePeriod=2 Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.774966 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.775205 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.840304 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:29 crc kubenswrapper[4989]: I1006 09:24:29.920849 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.011857 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.107175 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6p2r\" (UniqueName: \"kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r\") pod \"ee84970d-edde-4e4a-87a1-d75fda40b295\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.107252 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content\") pod \"ee84970d-edde-4e4a-87a1-d75fda40b295\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.107483 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities\") pod \"ee84970d-edde-4e4a-87a1-d75fda40b295\" (UID: \"ee84970d-edde-4e4a-87a1-d75fda40b295\") " Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.108556 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities" (OuterVolumeSpecName: "utilities") pod "ee84970d-edde-4e4a-87a1-d75fda40b295" (UID: "ee84970d-edde-4e4a-87a1-d75fda40b295"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.117570 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r" (OuterVolumeSpecName: "kube-api-access-b6p2r") pod "ee84970d-edde-4e4a-87a1-d75fda40b295" (UID: "ee84970d-edde-4e4a-87a1-d75fda40b295"). InnerVolumeSpecName "kube-api-access-b6p2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.127949 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee84970d-edde-4e4a-87a1-d75fda40b295" (UID: "ee84970d-edde-4e4a-87a1-d75fda40b295"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.209390 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.209432 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6p2r\" (UniqueName: \"kubernetes.io/projected/ee84970d-edde-4e4a-87a1-d75fda40b295-kube-api-access-b6p2r\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.209445 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee84970d-edde-4e4a-87a1-d75fda40b295-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.545131 4989 generic.go:334] "Generic (PLEG): container finished" podID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerID="a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b" exitCode=0 Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.545201 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerDied","Data":"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b"} Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.545250 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-stv6t" event={"ID":"ee84970d-edde-4e4a-87a1-d75fda40b295","Type":"ContainerDied","Data":"508d5d146e1104577a87d4a54d60d52618641b9f421afca866447fcf0c13353d"} Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.545275 4989 scope.go:117] "RemoveContainer" containerID="a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.545132 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-stv6t" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.546684 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t5j6x" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="registry-server" containerID="cri-o://6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806" gracePeriod=2 Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.580834 4989 scope.go:117] "RemoveContainer" containerID="06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.586901 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.591437 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-stv6t"] Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.608929 4989 scope.go:117] "RemoveContainer" containerID="de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.609007 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.720250 4989 scope.go:117] "RemoveContainer" containerID="a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b" Oct 06 09:24:30 crc kubenswrapper[4989]: E1006 09:24:30.720729 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b\": container with ID starting with a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b not found: ID does not exist" containerID="a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.720784 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b"} err="failed to get container status \"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b\": rpc error: code = NotFound desc = could not find container \"a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b\": container with ID starting with a1f1c22411f5a982abf15160a2acb40fd417e95e81f46328b5fed92fc6b7e77b not found: ID does not exist" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.720818 4989 scope.go:117] "RemoveContainer" containerID="06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5" Oct 06 09:24:30 crc kubenswrapper[4989]: E1006 09:24:30.721132 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5\": container with ID starting with 06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5 not found: ID does not exist" containerID="06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.721170 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5"} err="failed to get container status \"06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5\": rpc error: code = NotFound desc = could not find container \"06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5\": container with ID starting with 06a83a57b88fbae864ef11a00d5acc1ea95a14e31d621b52f326123f5071b6a5 not found: ID does not exist" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.721195 4989 scope.go:117] "RemoveContainer" containerID="de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904" Oct 06 09:24:30 crc kubenswrapper[4989]: E1006 09:24:30.721449 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904\": container with ID starting with de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904 not found: ID does not exist" containerID="de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.721477 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904"} err="failed to get container status \"de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904\": rpc error: code = NotFound desc = could not find container \"de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904\": container with ID starting with de4586c7b2f2612efea9e3676d7c9686839095d898ce7a2bdefd1f70323f0904 not found: ID does not exist" Oct 06 09:24:30 crc kubenswrapper[4989]: I1006 09:24:30.900867 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.020286 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities\") pod \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.020411 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content\") pod \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.020451 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qvdz\" (UniqueName: \"kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz\") pod \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\" (UID: \"e78cfbdd-cad3-46e5-9c56-e64e2595adfa\") " Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.021534 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities" (OuterVolumeSpecName: "utilities") pod "e78cfbdd-cad3-46e5-9c56-e64e2595adfa" (UID: "e78cfbdd-cad3-46e5-9c56-e64e2595adfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.028983 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz" (OuterVolumeSpecName: "kube-api-access-2qvdz") pod "e78cfbdd-cad3-46e5-9c56-e64e2595adfa" (UID: "e78cfbdd-cad3-46e5-9c56-e64e2595adfa"). InnerVolumeSpecName "kube-api-access-2qvdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.071428 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e78cfbdd-cad3-46e5-9c56-e64e2595adfa" (UID: "e78cfbdd-cad3-46e5-9c56-e64e2595adfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.123385 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.123419 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qvdz\" (UniqueName: \"kubernetes.io/projected/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-kube-api-access-2qvdz\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.123431 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78cfbdd-cad3-46e5-9c56-e64e2595adfa-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.556585 4989 generic.go:334] "Generic (PLEG): container finished" podID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerID="6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806" exitCode=0 Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.556799 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerDied","Data":"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806"} Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.556888 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t5j6x" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.556915 4989 scope.go:117] "RemoveContainer" containerID="6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.556903 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t5j6x" event={"ID":"e78cfbdd-cad3-46e5-9c56-e64e2595adfa","Type":"ContainerDied","Data":"96bc92cdeb000a1799304aae4a2946a45f5946f6681bcc00c2ac90cec302c3a6"} Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.590959 4989 scope.go:117] "RemoveContainer" containerID="f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.593880 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.604024 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t5j6x"] Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.628132 4989 scope.go:117] "RemoveContainer" containerID="23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.643614 4989 scope.go:117] "RemoveContainer" containerID="6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806" Oct 06 09:24:31 crc kubenswrapper[4989]: E1006 09:24:31.644158 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806\": container with ID starting with 6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806 not found: ID does not exist" containerID="6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.644186 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806"} err="failed to get container status \"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806\": rpc error: code = NotFound desc = could not find container \"6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806\": container with ID starting with 6f52b7d8fede1434668b4dffba36b8e78fd1bf359750c70751bc108059c44806 not found: ID does not exist" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.644206 4989 scope.go:117] "RemoveContainer" containerID="f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd" Oct 06 09:24:31 crc kubenswrapper[4989]: E1006 09:24:31.644690 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd\": container with ID starting with f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd not found: ID does not exist" containerID="f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.644732 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd"} err="failed to get container status \"f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd\": rpc error: code = NotFound desc = could not find container \"f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd\": container with ID starting with f4a92ca6b6574a0ab03697242344e1498c869905e9ca52a20ece4d2374296bcd not found: ID does not exist" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.644760 4989 scope.go:117] "RemoveContainer" containerID="23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08" Oct 06 09:24:31 crc kubenswrapper[4989]: E1006 09:24:31.645011 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08\": container with ID starting with 23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08 not found: ID does not exist" containerID="23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.645100 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08"} err="failed to get container status \"23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08\": rpc error: code = NotFound desc = could not find container \"23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08\": container with ID starting with 23563ae88e5e49200a08736818b1ab91f6845855acb154c756965b0fc8752a08 not found: ID does not exist" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.952420 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" path="/var/lib/kubelet/pods/e78cfbdd-cad3-46e5-9c56-e64e2595adfa/volumes" Oct 06 09:24:31 crc kubenswrapper[4989]: I1006 09:24:31.953452 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" path="/var/lib/kubelet/pods/ee84970d-edde-4e4a-87a1-d75fda40b295/volumes" Oct 06 09:24:32 crc kubenswrapper[4989]: I1006 09:24:32.410991 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:33 crc kubenswrapper[4989]: I1006 09:24:33.577109 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6mgp7" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="registry-server" containerID="cri-o://debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947" gracePeriod=2 Oct 06 09:24:33 crc kubenswrapper[4989]: I1006 09:24:33.996514 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.166336 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities\") pod \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.166410 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content\") pod \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.166553 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx2wm\" (UniqueName: \"kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm\") pod \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\" (UID: \"4e8e26d6-e46a-4c4d-bf50-45da05934bbc\") " Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.167216 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities" (OuterVolumeSpecName: "utilities") pod "4e8e26d6-e46a-4c4d-bf50-45da05934bbc" (UID: "4e8e26d6-e46a-4c4d-bf50-45da05934bbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.172403 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm" (OuterVolumeSpecName: "kube-api-access-kx2wm") pod "4e8e26d6-e46a-4c4d-bf50-45da05934bbc" (UID: "4e8e26d6-e46a-4c4d-bf50-45da05934bbc"). InnerVolumeSpecName "kube-api-access-kx2wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.211201 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e8e26d6-e46a-4c4d-bf50-45da05934bbc" (UID: "4e8e26d6-e46a-4c4d-bf50-45da05934bbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.267867 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx2wm\" (UniqueName: \"kubernetes.io/projected/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-kube-api-access-kx2wm\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.267904 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.267913 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8e26d6-e46a-4c4d-bf50-45da05934bbc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.588935 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mgp7" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.589111 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerDied","Data":"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947"} Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.589958 4989 scope.go:117] "RemoveContainer" containerID="debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.588827 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerID="debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947" exitCode=0 Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.590060 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mgp7" event={"ID":"4e8e26d6-e46a-4c4d-bf50-45da05934bbc","Type":"ContainerDied","Data":"da53cbb078319736db75c04588d11df00dc40a9022d759e046d58a1f84808797"} Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.609979 4989 scope.go:117] "RemoveContainer" containerID="8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.623973 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.633757 4989 scope.go:117] "RemoveContainer" containerID="66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.636324 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6mgp7"] Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.649672 4989 scope.go:117] "RemoveContainer" containerID="debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947" Oct 06 09:24:34 crc kubenswrapper[4989]: E1006 09:24:34.650088 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947\": container with ID starting with debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947 not found: ID does not exist" containerID="debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.650127 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947"} err="failed to get container status \"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947\": rpc error: code = NotFound desc = could not find container \"debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947\": container with ID starting with debf4c827309797384fb36839f9673f1c71cfe931f70d2110be3df468a23d947 not found: ID does not exist" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.650150 4989 scope.go:117] "RemoveContainer" containerID="8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205" Oct 06 09:24:34 crc kubenswrapper[4989]: E1006 09:24:34.650402 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205\": container with ID starting with 8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205 not found: ID does not exist" containerID="8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.650501 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205"} err="failed to get container status \"8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205\": rpc error: code = NotFound desc = could not find container \"8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205\": container with ID starting with 8ec1ea397e12fa1bbad863fb2f452334fc78fba2bde274f93921682f455f5205 not found: ID does not exist" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.650529 4989 scope.go:117] "RemoveContainer" containerID="66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6" Oct 06 09:24:34 crc kubenswrapper[4989]: E1006 09:24:34.650838 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6\": container with ID starting with 66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6 not found: ID does not exist" containerID="66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6" Oct 06 09:24:34 crc kubenswrapper[4989]: I1006 09:24:34.650862 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6"} err="failed to get container status \"66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6\": rpc error: code = NotFound desc = could not find container \"66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6\": container with ID starting with 66f4969123650e6798b354fbc1bca881996ccec10fa074b90f928580f2cbbed6 not found: ID does not exist" Oct 06 09:24:35 crc kubenswrapper[4989]: I1006 09:24:35.935848 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:24:35 crc kubenswrapper[4989]: I1006 09:24:35.951132 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" path="/var/lib/kubelet/pods/4e8e26d6-e46a-4c4d-bf50-45da05934bbc/volumes" Oct 06 09:24:36 crc kubenswrapper[4989]: I1006 09:24:36.622630 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04"} Oct 06 09:27:03 crc kubenswrapper[4989]: I1006 09:27:03.935825 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:27:03 crc kubenswrapper[4989]: I1006 09:27:03.936708 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:27:33 crc kubenswrapper[4989]: I1006 09:27:33.936359 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:27:33 crc kubenswrapper[4989]: I1006 09:27:33.936988 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.210810 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212289 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212421 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212470 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212480 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212498 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212508 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212524 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212531 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212544 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212552 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212562 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212573 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="extract-content" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212589 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212596 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212667 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212677 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="extract-utilities" Oct 06 09:27:46 crc kubenswrapper[4989]: E1006 09:27:46.212692 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212698 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212907 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8e26d6-e46a-4c4d-bf50-45da05934bbc" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212933 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee84970d-edde-4e4a-87a1-d75fda40b295" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.212951 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78cfbdd-cad3-46e5-9c56-e64e2595adfa" containerName="registry-server" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.214704 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.225718 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.307123 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.307205 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k54kb\" (UniqueName: \"kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.307257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.408792 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.408976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k54kb\" (UniqueName: \"kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.409099 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.409563 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.409751 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.430219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k54kb\" (UniqueName: \"kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb\") pod \"redhat-operators-q8zxl\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.535680 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:46 crc kubenswrapper[4989]: I1006 09:27:46.976106 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:27:47 crc kubenswrapper[4989]: I1006 09:27:47.299415 4989 generic.go:334] "Generic (PLEG): container finished" podID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerID="b2b99446fe58f705cf1d7bf40f3475dcf8b5e99bfe5b08f6bc0c89f4fddaa2db" exitCode=0 Oct 06 09:27:47 crc kubenswrapper[4989]: I1006 09:27:47.299473 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerDied","Data":"b2b99446fe58f705cf1d7bf40f3475dcf8b5e99bfe5b08f6bc0c89f4fddaa2db"} Oct 06 09:27:47 crc kubenswrapper[4989]: I1006 09:27:47.299533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerStarted","Data":"84c02fcd1acb0b6a5128b8666c5901eac0d080879e44806c48fd17be6692ec6f"} Oct 06 09:27:48 crc kubenswrapper[4989]: I1006 09:27:48.308878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerStarted","Data":"17d9b6b16108e1574657e071eac9cc7759b373f0db7911367eea48c9105c8d7c"} Oct 06 09:27:49 crc kubenswrapper[4989]: I1006 09:27:49.322852 4989 generic.go:334] "Generic (PLEG): container finished" podID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerID="17d9b6b16108e1574657e071eac9cc7759b373f0db7911367eea48c9105c8d7c" exitCode=0 Oct 06 09:27:49 crc kubenswrapper[4989]: I1006 09:27:49.323534 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerDied","Data":"17d9b6b16108e1574657e071eac9cc7759b373f0db7911367eea48c9105c8d7c"} Oct 06 09:27:50 crc kubenswrapper[4989]: I1006 09:27:50.333923 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerStarted","Data":"340f996cf53b8693437cc118be0ce04782d4697d7d556ab4f847e4ce04ad1995"} Oct 06 09:27:50 crc kubenswrapper[4989]: I1006 09:27:50.355208 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q8zxl" podStartSLOduration=1.910036996 podStartE2EDuration="4.355192408s" podCreationTimestamp="2025-10-06 09:27:46 +0000 UTC" firstStartedPulling="2025-10-06 09:27:47.303032238 +0000 UTC m=+2918.093057818" lastFinishedPulling="2025-10-06 09:27:49.74818762 +0000 UTC m=+2920.538213230" observedRunningTime="2025-10-06 09:27:50.351786239 +0000 UTC m=+2921.141811859" watchObservedRunningTime="2025-10-06 09:27:50.355192408 +0000 UTC m=+2921.145217988" Oct 06 09:27:56 crc kubenswrapper[4989]: I1006 09:27:56.536286 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:56 crc kubenswrapper[4989]: I1006 09:27:56.536614 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:56 crc kubenswrapper[4989]: I1006 09:27:56.610182 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:57 crc kubenswrapper[4989]: I1006 09:27:57.476514 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:27:57 crc kubenswrapper[4989]: I1006 09:27:57.595671 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:27:59 crc kubenswrapper[4989]: I1006 09:27:59.434399 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q8zxl" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="registry-server" containerID="cri-o://340f996cf53b8693437cc118be0ce04782d4697d7d556ab4f847e4ce04ad1995" gracePeriod=2 Oct 06 09:28:02 crc kubenswrapper[4989]: I1006 09:28:02.474793 4989 generic.go:334] "Generic (PLEG): container finished" podID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerID="340f996cf53b8693437cc118be0ce04782d4697d7d556ab4f847e4ce04ad1995" exitCode=0 Oct 06 09:28:02 crc kubenswrapper[4989]: I1006 09:28:02.474863 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerDied","Data":"340f996cf53b8693437cc118be0ce04782d4697d7d556ab4f847e4ce04ad1995"} Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.090377 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.180638 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content\") pod \"218eaa3f-607e-4313-8d99-feadb9d61f75\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.180738 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k54kb\" (UniqueName: \"kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb\") pod \"218eaa3f-607e-4313-8d99-feadb9d61f75\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.180772 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities\") pod \"218eaa3f-607e-4313-8d99-feadb9d61f75\" (UID: \"218eaa3f-607e-4313-8d99-feadb9d61f75\") " Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.181793 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities" (OuterVolumeSpecName: "utilities") pod "218eaa3f-607e-4313-8d99-feadb9d61f75" (UID: "218eaa3f-607e-4313-8d99-feadb9d61f75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.186945 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb" (OuterVolumeSpecName: "kube-api-access-k54kb") pod "218eaa3f-607e-4313-8d99-feadb9d61f75" (UID: "218eaa3f-607e-4313-8d99-feadb9d61f75"). InnerVolumeSpecName "kube-api-access-k54kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.263943 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "218eaa3f-607e-4313-8d99-feadb9d61f75" (UID: "218eaa3f-607e-4313-8d99-feadb9d61f75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.282393 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k54kb\" (UniqueName: \"kubernetes.io/projected/218eaa3f-607e-4313-8d99-feadb9d61f75-kube-api-access-k54kb\") on node \"crc\" DevicePath \"\"" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.282435 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.282452 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218eaa3f-607e-4313-8d99-feadb9d61f75-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.483560 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q8zxl" event={"ID":"218eaa3f-607e-4313-8d99-feadb9d61f75","Type":"ContainerDied","Data":"84c02fcd1acb0b6a5128b8666c5901eac0d080879e44806c48fd17be6692ec6f"} Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.483613 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q8zxl" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.484472 4989 scope.go:117] "RemoveContainer" containerID="340f996cf53b8693437cc118be0ce04782d4697d7d556ab4f847e4ce04ad1995" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.526537 4989 scope.go:117] "RemoveContainer" containerID="17d9b6b16108e1574657e071eac9cc7759b373f0db7911367eea48c9105c8d7c" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.533561 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.538918 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q8zxl"] Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.575037 4989 scope.go:117] "RemoveContainer" containerID="b2b99446fe58f705cf1d7bf40f3475dcf8b5e99bfe5b08f6bc0c89f4fddaa2db" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.936946 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.937484 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.949281 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" path="/var/lib/kubelet/pods/218eaa3f-607e-4313-8d99-feadb9d61f75/volumes" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.950584 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.951460 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:28:03 crc kubenswrapper[4989]: I1006 09:28:03.952269 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04" gracePeriod=600 Oct 06 09:28:04 crc kubenswrapper[4989]: I1006 09:28:04.493775 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04" exitCode=0 Oct 06 09:28:04 crc kubenswrapper[4989]: I1006 09:28:04.493849 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04"} Oct 06 09:28:04 crc kubenswrapper[4989]: I1006 09:28:04.493946 4989 scope.go:117] "RemoveContainer" containerID="afc0eafb7ba72161263387f13a53e106154c02d96a38df1915f5b3ca74eefc98" Oct 06 09:28:05 crc kubenswrapper[4989]: I1006 09:28:05.505099 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0"} Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.181466 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd"] Oct 06 09:30:00 crc kubenswrapper[4989]: E1006 09:30:00.182487 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="extract-content" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.182509 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="extract-content" Oct 06 09:30:00 crc kubenswrapper[4989]: E1006 09:30:00.182528 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="extract-utilities" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.182537 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="extract-utilities" Oct 06 09:30:00 crc kubenswrapper[4989]: E1006 09:30:00.182569 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="registry-server" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.182578 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="registry-server" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.182813 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="218eaa3f-607e-4313-8d99-feadb9d61f75" containerName="registry-server" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.183494 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.185917 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.186412 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.230456 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd"] Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.287846 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4qwb\" (UniqueName: \"kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.287904 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.287933 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.389526 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.389569 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.389650 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4qwb\" (UniqueName: \"kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.390533 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.396435 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.405094 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4qwb\" (UniqueName: \"kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb\") pod \"collect-profiles-29329050-prdpd\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.501381 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:00 crc kubenswrapper[4989]: I1006 09:30:00.996591 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd"] Oct 06 09:30:01 crc kubenswrapper[4989]: I1006 09:30:01.520872 4989 generic.go:334] "Generic (PLEG): container finished" podID="6e3ed952-b113-406d-a267-22445a1e4ead" containerID="a01613d3f1d62886f1ef7d2fbf27fd2e59e90e015b2242e39fa2ba21357d1310" exitCode=0 Oct 06 09:30:01 crc kubenswrapper[4989]: I1006 09:30:01.520947 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" event={"ID":"6e3ed952-b113-406d-a267-22445a1e4ead","Type":"ContainerDied","Data":"a01613d3f1d62886f1ef7d2fbf27fd2e59e90e015b2242e39fa2ba21357d1310"} Oct 06 09:30:01 crc kubenswrapper[4989]: I1006 09:30:01.520991 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" event={"ID":"6e3ed952-b113-406d-a267-22445a1e4ead","Type":"ContainerStarted","Data":"c92f3f87be414e1a3246bb8fd0eb179ca54b66cf88b6b4b55af052e939ee1586"} Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.797380 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.823056 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume\") pod \"6e3ed952-b113-406d-a267-22445a1e4ead\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.823147 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4qwb\" (UniqueName: \"kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb\") pod \"6e3ed952-b113-406d-a267-22445a1e4ead\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.823207 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume\") pod \"6e3ed952-b113-406d-a267-22445a1e4ead\" (UID: \"6e3ed952-b113-406d-a267-22445a1e4ead\") " Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.823806 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume" (OuterVolumeSpecName: "config-volume") pod "6e3ed952-b113-406d-a267-22445a1e4ead" (UID: "6e3ed952-b113-406d-a267-22445a1e4ead"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.824235 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6e3ed952-b113-406d-a267-22445a1e4ead-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.830290 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb" (OuterVolumeSpecName: "kube-api-access-n4qwb") pod "6e3ed952-b113-406d-a267-22445a1e4ead" (UID: "6e3ed952-b113-406d-a267-22445a1e4ead"). InnerVolumeSpecName "kube-api-access-n4qwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.830738 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6e3ed952-b113-406d-a267-22445a1e4ead" (UID: "6e3ed952-b113-406d-a267-22445a1e4ead"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.925103 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4qwb\" (UniqueName: \"kubernetes.io/projected/6e3ed952-b113-406d-a267-22445a1e4ead-kube-api-access-n4qwb\") on node \"crc\" DevicePath \"\"" Oct 06 09:30:02 crc kubenswrapper[4989]: I1006 09:30:02.925135 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6e3ed952-b113-406d-a267-22445a1e4ead-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.537846 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.537732 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd" event={"ID":"6e3ed952-b113-406d-a267-22445a1e4ead","Type":"ContainerDied","Data":"c92f3f87be414e1a3246bb8fd0eb179ca54b66cf88b6b4b55af052e939ee1586"} Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.538984 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c92f3f87be414e1a3246bb8fd0eb179ca54b66cf88b6b4b55af052e939ee1586" Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.862958 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc"] Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.867828 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329005-vjcgc"] Oct 06 09:30:03 crc kubenswrapper[4989]: I1006 09:30:03.943064 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dcddcf3-b68d-4327-b441-26cfac414aab" path="/var/lib/kubelet/pods/5dcddcf3-b68d-4327-b441-26cfac414aab/volumes" Oct 06 09:30:19 crc kubenswrapper[4989]: I1006 09:30:19.053085 4989 scope.go:117] "RemoveContainer" containerID="53ed129a133edb823a9ffcd57f0b3f6e401d72815cdd8ae1f390abea11220dde" Oct 06 09:30:33 crc kubenswrapper[4989]: I1006 09:30:33.936295 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:30:33 crc kubenswrapper[4989]: I1006 09:30:33.938088 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:31:03 crc kubenswrapper[4989]: I1006 09:31:03.935865 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:31:03 crc kubenswrapper[4989]: I1006 09:31:03.944804 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:31:33 crc kubenswrapper[4989]: I1006 09:31:33.936312 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:31:33 crc kubenswrapper[4989]: I1006 09:31:33.937002 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:31:33 crc kubenswrapper[4989]: I1006 09:31:33.950445 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:31:33 crc kubenswrapper[4989]: I1006 09:31:33.951215 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:31:33 crc kubenswrapper[4989]: I1006 09:31:33.951315 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" gracePeriod=600 Oct 06 09:31:34 crc kubenswrapper[4989]: I1006 09:31:34.357536 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" exitCode=0 Oct 06 09:31:34 crc kubenswrapper[4989]: I1006 09:31:34.357613 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0"} Oct 06 09:31:34 crc kubenswrapper[4989]: I1006 09:31:34.357707 4989 scope.go:117] "RemoveContainer" containerID="740977e1d6a41d0c952521066703336cfdee4d2a89af1576aa054b41a2106f04" Oct 06 09:31:34 crc kubenswrapper[4989]: E1006 09:31:34.605851 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:31:35 crc kubenswrapper[4989]: I1006 09:31:35.370750 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:31:35 crc kubenswrapper[4989]: E1006 09:31:35.371634 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:31:45 crc kubenswrapper[4989]: I1006 09:31:45.936487 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:31:45 crc kubenswrapper[4989]: E1006 09:31:45.937288 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:31:59 crc kubenswrapper[4989]: I1006 09:31:59.942933 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:31:59 crc kubenswrapper[4989]: E1006 09:31:59.943942 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:32:13 crc kubenswrapper[4989]: I1006 09:32:13.936779 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:32:13 crc kubenswrapper[4989]: E1006 09:32:13.938187 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:32:26 crc kubenswrapper[4989]: I1006 09:32:26.936850 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:32:26 crc kubenswrapper[4989]: E1006 09:32:26.937590 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:32:41 crc kubenswrapper[4989]: I1006 09:32:41.936433 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:32:41 crc kubenswrapper[4989]: E1006 09:32:41.937435 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:32:55 crc kubenswrapper[4989]: I1006 09:32:55.936298 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:32:55 crc kubenswrapper[4989]: E1006 09:32:55.937212 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:33:09 crc kubenswrapper[4989]: I1006 09:33:09.945864 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:33:09 crc kubenswrapper[4989]: E1006 09:33:09.948203 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:33:21 crc kubenswrapper[4989]: I1006 09:33:21.935930 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:33:21 crc kubenswrapper[4989]: E1006 09:33:21.938528 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:33:36 crc kubenswrapper[4989]: I1006 09:33:36.936924 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:33:36 crc kubenswrapper[4989]: E1006 09:33:36.938033 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:33:48 crc kubenswrapper[4989]: I1006 09:33:48.937782 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:33:48 crc kubenswrapper[4989]: E1006 09:33:48.939325 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:33:59 crc kubenswrapper[4989]: I1006 09:33:59.947925 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:33:59 crc kubenswrapper[4989]: E1006 09:33:59.949061 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:34:13 crc kubenswrapper[4989]: I1006 09:34:13.936609 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:34:13 crc kubenswrapper[4989]: E1006 09:34:13.937446 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:34:27 crc kubenswrapper[4989]: I1006 09:34:27.935947 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:34:27 crc kubenswrapper[4989]: E1006 09:34:27.936581 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.765583 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:41 crc kubenswrapper[4989]: E1006 09:34:41.766761 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e3ed952-b113-406d-a267-22445a1e4ead" containerName="collect-profiles" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.766779 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e3ed952-b113-406d-a267-22445a1e4ead" containerName="collect-profiles" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.767019 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e3ed952-b113-406d-a267-22445a1e4ead" containerName="collect-profiles" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.768434 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.777678 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.866017 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.866127 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j47vn\" (UniqueName: \"kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.866150 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.968288 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j47vn\" (UniqueName: \"kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.968372 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.968451 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.969024 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.969215 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:41 crc kubenswrapper[4989]: I1006 09:34:41.993134 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j47vn\" (UniqueName: \"kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn\") pod \"certified-operators-pfkb6\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.096815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.374023 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.936054 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:34:42 crc kubenswrapper[4989]: E1006 09:34:42.937771 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.968889 4989 generic.go:334] "Generic (PLEG): container finished" podID="eaf8aa91-198a-4e34-8376-528711fec688" containerID="e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512" exitCode=0 Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.968935 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerDied","Data":"e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512"} Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.968961 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerStarted","Data":"571da6fbb997543026b9334f8e40f05f8c7928ff7b8fbb4f52de2fccfdff6bfb"} Oct 06 09:34:42 crc kubenswrapper[4989]: I1006 09:34:42.971893 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:34:44 crc kubenswrapper[4989]: I1006 09:34:44.982275 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerStarted","Data":"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575"} Oct 06 09:34:45 crc kubenswrapper[4989]: I1006 09:34:45.993406 4989 generic.go:334] "Generic (PLEG): container finished" podID="eaf8aa91-198a-4e34-8376-528711fec688" containerID="2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575" exitCode=0 Oct 06 09:34:45 crc kubenswrapper[4989]: I1006 09:34:45.993463 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerDied","Data":"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575"} Oct 06 09:34:47 crc kubenswrapper[4989]: I1006 09:34:47.006867 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerStarted","Data":"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277"} Oct 06 09:34:47 crc kubenswrapper[4989]: I1006 09:34:47.055294 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pfkb6" podStartSLOduration=2.601301491 podStartE2EDuration="6.055260915s" podCreationTimestamp="2025-10-06 09:34:41 +0000 UTC" firstStartedPulling="2025-10-06 09:34:42.97141246 +0000 UTC m=+3333.761438070" lastFinishedPulling="2025-10-06 09:34:46.425371914 +0000 UTC m=+3337.215397494" observedRunningTime="2025-10-06 09:34:47.051380765 +0000 UTC m=+3337.841406345" watchObservedRunningTime="2025-10-06 09:34:47.055260915 +0000 UTC m=+3337.845286535" Oct 06 09:34:52 crc kubenswrapper[4989]: I1006 09:34:52.097601 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:52 crc kubenswrapper[4989]: I1006 09:34:52.097988 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:52 crc kubenswrapper[4989]: I1006 09:34:52.160050 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:53 crc kubenswrapper[4989]: I1006 09:34:53.121178 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:53 crc kubenswrapper[4989]: I1006 09:34:53.194258 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.804562 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.807730 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.818082 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.862561 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmqqv\" (UniqueName: \"kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.862806 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.862916 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.936247 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:34:54 crc kubenswrapper[4989]: E1006 09:34:54.936795 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.964102 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmqqv\" (UniqueName: \"kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.964175 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.964213 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.964594 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.964612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:54 crc kubenswrapper[4989]: I1006 09:34:54.984851 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmqqv\" (UniqueName: \"kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv\") pod \"redhat-marketplace-p8j7k\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.083746 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pfkb6" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="registry-server" containerID="cri-o://4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277" gracePeriod=2 Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.142984 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.537484 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.595847 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:34:55 crc kubenswrapper[4989]: W1006 09:34:55.598844 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod990f7f44_e7e1_4a54_a0ee_ec690b52cca9.slice/crio-e61695694888530abc29acdad5eedfe9ea42c1228ed149db49bc773cad6ceead WatchSource:0}: Error finding container e61695694888530abc29acdad5eedfe9ea42c1228ed149db49bc773cad6ceead: Status 404 returned error can't find the container with id e61695694888530abc29acdad5eedfe9ea42c1228ed149db49bc773cad6ceead Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.675251 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content\") pod \"eaf8aa91-198a-4e34-8376-528711fec688\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.675553 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j47vn\" (UniqueName: \"kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn\") pod \"eaf8aa91-198a-4e34-8376-528711fec688\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.675789 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities\") pod \"eaf8aa91-198a-4e34-8376-528711fec688\" (UID: \"eaf8aa91-198a-4e34-8376-528711fec688\") " Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.677584 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities" (OuterVolumeSpecName: "utilities") pod "eaf8aa91-198a-4e34-8376-528711fec688" (UID: "eaf8aa91-198a-4e34-8376-528711fec688"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.683583 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn" (OuterVolumeSpecName: "kube-api-access-j47vn") pod "eaf8aa91-198a-4e34-8376-528711fec688" (UID: "eaf8aa91-198a-4e34-8376-528711fec688"). InnerVolumeSpecName "kube-api-access-j47vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.749878 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eaf8aa91-198a-4e34-8376-528711fec688" (UID: "eaf8aa91-198a-4e34-8376-528711fec688"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.777391 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.777428 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf8aa91-198a-4e34-8376-528711fec688-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:34:55 crc kubenswrapper[4989]: I1006 09:34:55.777442 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j47vn\" (UniqueName: \"kubernetes.io/projected/eaf8aa91-198a-4e34-8376-528711fec688-kube-api-access-j47vn\") on node \"crc\" DevicePath \"\"" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.100587 4989 generic.go:334] "Generic (PLEG): container finished" podID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerID="a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10" exitCode=0 Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.100919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerDied","Data":"a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10"} Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.100945 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerStarted","Data":"e61695694888530abc29acdad5eedfe9ea42c1228ed149db49bc773cad6ceead"} Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.104013 4989 generic.go:334] "Generic (PLEG): container finished" podID="eaf8aa91-198a-4e34-8376-528711fec688" containerID="4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277" exitCode=0 Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.104045 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerDied","Data":"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277"} Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.104060 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfkb6" event={"ID":"eaf8aa91-198a-4e34-8376-528711fec688","Type":"ContainerDied","Data":"571da6fbb997543026b9334f8e40f05f8c7928ff7b8fbb4f52de2fccfdff6bfb"} Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.104076 4989 scope.go:117] "RemoveContainer" containerID="4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.104174 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfkb6" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.131899 4989 scope.go:117] "RemoveContainer" containerID="2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.166356 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.172486 4989 scope.go:117] "RemoveContainer" containerID="e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.177026 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pfkb6"] Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.196251 4989 scope.go:117] "RemoveContainer" containerID="4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277" Oct 06 09:34:56 crc kubenswrapper[4989]: E1006 09:34:56.197165 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277\": container with ID starting with 4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277 not found: ID does not exist" containerID="4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.197229 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277"} err="failed to get container status \"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277\": rpc error: code = NotFound desc = could not find container \"4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277\": container with ID starting with 4aa4b3dd34cf12b9e5adac6094a8c535615ff73b2986a8489eea397fca8fe277 not found: ID does not exist" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.197264 4989 scope.go:117] "RemoveContainer" containerID="2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575" Oct 06 09:34:56 crc kubenswrapper[4989]: E1006 09:34:56.197884 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575\": container with ID starting with 2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575 not found: ID does not exist" containerID="2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.197931 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575"} err="failed to get container status \"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575\": rpc error: code = NotFound desc = could not find container \"2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575\": container with ID starting with 2826656385fbed54ae11812aef64df59a5d9addb09677c133dd33e0aa6108575 not found: ID does not exist" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.197959 4989 scope.go:117] "RemoveContainer" containerID="e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512" Oct 06 09:34:56 crc kubenswrapper[4989]: E1006 09:34:56.198529 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512\": container with ID starting with e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512 not found: ID does not exist" containerID="e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512" Oct 06 09:34:56 crc kubenswrapper[4989]: I1006 09:34:56.198563 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512"} err="failed to get container status \"e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512\": rpc error: code = NotFound desc = could not find container \"e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512\": container with ID starting with e3a89401651dbfa0d2bef6d4a71c773f5685ee2bdf53e8e0e32237b42ec8f512 not found: ID does not exist" Oct 06 09:34:57 crc kubenswrapper[4989]: I1006 09:34:57.116155 4989 generic.go:334] "Generic (PLEG): container finished" podID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerID="2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279" exitCode=0 Oct 06 09:34:57 crc kubenswrapper[4989]: I1006 09:34:57.116253 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerDied","Data":"2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279"} Oct 06 09:34:57 crc kubenswrapper[4989]: I1006 09:34:57.946550 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf8aa91-198a-4e34-8376-528711fec688" path="/var/lib/kubelet/pods/eaf8aa91-198a-4e34-8376-528711fec688/volumes" Oct 06 09:34:58 crc kubenswrapper[4989]: I1006 09:34:58.138806 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerStarted","Data":"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255"} Oct 06 09:34:58 crc kubenswrapper[4989]: I1006 09:34:58.160859 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p8j7k" podStartSLOduration=2.701618934 podStartE2EDuration="4.160842518s" podCreationTimestamp="2025-10-06 09:34:54 +0000 UTC" firstStartedPulling="2025-10-06 09:34:56.102543899 +0000 UTC m=+3346.892569479" lastFinishedPulling="2025-10-06 09:34:57.561767443 +0000 UTC m=+3348.351793063" observedRunningTime="2025-10-06 09:34:58.155405503 +0000 UTC m=+3348.945431083" watchObservedRunningTime="2025-10-06 09:34:58.160842518 +0000 UTC m=+3348.950868098" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.398331 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:02 crc kubenswrapper[4989]: E1006 09:35:02.398959 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="registry-server" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.398971 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="registry-server" Oct 06 09:35:02 crc kubenswrapper[4989]: E1006 09:35:02.398980 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="extract-content" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.398987 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="extract-content" Oct 06 09:35:02 crc kubenswrapper[4989]: E1006 09:35:02.398999 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="extract-utilities" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.399005 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="extract-utilities" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.399124 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf8aa91-198a-4e34-8376-528711fec688" containerName="registry-server" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.400104 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.430428 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.478637 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.478745 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75gtq\" (UniqueName: \"kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.478808 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.579973 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.580042 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.580081 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75gtq\" (UniqueName: \"kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.580586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.580601 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.603478 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75gtq\" (UniqueName: \"kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq\") pod \"community-operators-ml5rl\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:02 crc kubenswrapper[4989]: I1006 09:35:02.730639 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:03 crc kubenswrapper[4989]: I1006 09:35:03.035267 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:03 crc kubenswrapper[4989]: I1006 09:35:03.188629 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerStarted","Data":"08f94091df92461683c05a2c8de45e2b05b17f93381f1875a83d613724300bf2"} Oct 06 09:35:04 crc kubenswrapper[4989]: I1006 09:35:04.199965 4989 generic.go:334] "Generic (PLEG): container finished" podID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerID="4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919" exitCode=0 Oct 06 09:35:04 crc kubenswrapper[4989]: I1006 09:35:04.200042 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerDied","Data":"4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919"} Oct 06 09:35:05 crc kubenswrapper[4989]: I1006 09:35:05.144536 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:05 crc kubenswrapper[4989]: I1006 09:35:05.144862 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:05 crc kubenswrapper[4989]: I1006 09:35:05.196164 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:05 crc kubenswrapper[4989]: I1006 09:35:05.251646 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:06 crc kubenswrapper[4989]: I1006 09:35:06.216116 4989 generic.go:334] "Generic (PLEG): container finished" podID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerID="31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31" exitCode=0 Oct 06 09:35:06 crc kubenswrapper[4989]: I1006 09:35:06.216176 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerDied","Data":"31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31"} Oct 06 09:35:07 crc kubenswrapper[4989]: I1006 09:35:07.225807 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerStarted","Data":"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32"} Oct 06 09:35:07 crc kubenswrapper[4989]: I1006 09:35:07.254084 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ml5rl" podStartSLOduration=2.79633601 podStartE2EDuration="5.25405742s" podCreationTimestamp="2025-10-06 09:35:02 +0000 UTC" firstStartedPulling="2025-10-06 09:35:04.201914666 +0000 UTC m=+3354.991940246" lastFinishedPulling="2025-10-06 09:35:06.659636086 +0000 UTC m=+3357.449661656" observedRunningTime="2025-10-06 09:35:07.245825195 +0000 UTC m=+3358.035850795" watchObservedRunningTime="2025-10-06 09:35:07.25405742 +0000 UTC m=+3358.044083040" Oct 06 09:35:07 crc kubenswrapper[4989]: I1006 09:35:07.597464 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:35:07 crc kubenswrapper[4989]: I1006 09:35:07.598030 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p8j7k" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="registry-server" containerID="cri-o://afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255" gracePeriod=2 Oct 06 09:35:07 crc kubenswrapper[4989]: I1006 09:35:07.936207 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:35:07 crc kubenswrapper[4989]: E1006 09:35:07.937292 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.030632 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.057087 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities\") pod \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.057135 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content\") pod \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.057298 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmqqv\" (UniqueName: \"kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv\") pod \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\" (UID: \"990f7f44-e7e1-4a54-a0ee-ec690b52cca9\") " Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.058461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities" (OuterVolumeSpecName: "utilities") pod "990f7f44-e7e1-4a54-a0ee-ec690b52cca9" (UID: "990f7f44-e7e1-4a54-a0ee-ec690b52cca9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.066847 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv" (OuterVolumeSpecName: "kube-api-access-qmqqv") pod "990f7f44-e7e1-4a54-a0ee-ec690b52cca9" (UID: "990f7f44-e7e1-4a54-a0ee-ec690b52cca9"). InnerVolumeSpecName "kube-api-access-qmqqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.074615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "990f7f44-e7e1-4a54-a0ee-ec690b52cca9" (UID: "990f7f44-e7e1-4a54-a0ee-ec690b52cca9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.159001 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.159046 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.159060 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmqqv\" (UniqueName: \"kubernetes.io/projected/990f7f44-e7e1-4a54-a0ee-ec690b52cca9-kube-api-access-qmqqv\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.235053 4989 generic.go:334] "Generic (PLEG): container finished" podID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerID="afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255" exitCode=0 Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.235833 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerDied","Data":"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255"} Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.235885 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p8j7k" event={"ID":"990f7f44-e7e1-4a54-a0ee-ec690b52cca9","Type":"ContainerDied","Data":"e61695694888530abc29acdad5eedfe9ea42c1228ed149db49bc773cad6ceead"} Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.235841 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p8j7k" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.235912 4989 scope.go:117] "RemoveContainer" containerID="afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.258575 4989 scope.go:117] "RemoveContainer" containerID="2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.281577 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.287823 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p8j7k"] Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.294921 4989 scope.go:117] "RemoveContainer" containerID="a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.325399 4989 scope.go:117] "RemoveContainer" containerID="afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255" Oct 06 09:35:08 crc kubenswrapper[4989]: E1006 09:35:08.325884 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255\": container with ID starting with afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255 not found: ID does not exist" containerID="afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.325930 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255"} err="failed to get container status \"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255\": rpc error: code = NotFound desc = could not find container \"afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255\": container with ID starting with afd383557bf381946aa71700e6f653c51ddccd3bf95ba1cbe03c4440d576c255 not found: ID does not exist" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.325958 4989 scope.go:117] "RemoveContainer" containerID="2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279" Oct 06 09:35:08 crc kubenswrapper[4989]: E1006 09:35:08.326297 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279\": container with ID starting with 2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279 not found: ID does not exist" containerID="2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.326321 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279"} err="failed to get container status \"2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279\": rpc error: code = NotFound desc = could not find container \"2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279\": container with ID starting with 2f4da1ae80a75b5adf278d4f9069f71af5e57970ba962cc580b2a6f404e95279 not found: ID does not exist" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.326338 4989 scope.go:117] "RemoveContainer" containerID="a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10" Oct 06 09:35:08 crc kubenswrapper[4989]: E1006 09:35:08.326928 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10\": container with ID starting with a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10 not found: ID does not exist" containerID="a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10" Oct 06 09:35:08 crc kubenswrapper[4989]: I1006 09:35:08.326953 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10"} err="failed to get container status \"a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10\": rpc error: code = NotFound desc = could not find container \"a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10\": container with ID starting with a7f461915fd0d97d65f15bddd0ba8f73ddaa5aa759b14afb5863260569ef3b10 not found: ID does not exist" Oct 06 09:35:09 crc kubenswrapper[4989]: I1006 09:35:09.948626 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" path="/var/lib/kubelet/pods/990f7f44-e7e1-4a54-a0ee-ec690b52cca9/volumes" Oct 06 09:35:12 crc kubenswrapper[4989]: I1006 09:35:12.731227 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:12 crc kubenswrapper[4989]: I1006 09:35:12.731608 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:12 crc kubenswrapper[4989]: I1006 09:35:12.794193 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:13 crc kubenswrapper[4989]: I1006 09:35:13.325319 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:14 crc kubenswrapper[4989]: I1006 09:35:14.395903 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.302251 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ml5rl" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="registry-server" containerID="cri-o://2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32" gracePeriod=2 Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.735550 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.788734 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities\") pod \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.788935 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75gtq\" (UniqueName: \"kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq\") pod \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.788957 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities" (OuterVolumeSpecName: "utilities") pod "7c9b2ea9-5417-471b-9e2d-d3e0df537269" (UID: "7c9b2ea9-5417-471b-9e2d-d3e0df537269"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.789046 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content\") pod \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\" (UID: \"7c9b2ea9-5417-471b-9e2d-d3e0df537269\") " Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.789519 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.794491 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq" (OuterVolumeSpecName: "kube-api-access-75gtq") pod "7c9b2ea9-5417-471b-9e2d-d3e0df537269" (UID: "7c9b2ea9-5417-471b-9e2d-d3e0df537269"). InnerVolumeSpecName "kube-api-access-75gtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.841668 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c9b2ea9-5417-471b-9e2d-d3e0df537269" (UID: "7c9b2ea9-5417-471b-9e2d-d3e0df537269"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.890986 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75gtq\" (UniqueName: \"kubernetes.io/projected/7c9b2ea9-5417-471b-9e2d-d3e0df537269-kube-api-access-75gtq\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:15 crc kubenswrapper[4989]: I1006 09:35:15.891029 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c9b2ea9-5417-471b-9e2d-d3e0df537269-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.313422 4989 generic.go:334] "Generic (PLEG): container finished" podID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerID="2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32" exitCode=0 Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.313508 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerDied","Data":"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32"} Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.313838 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ml5rl" event={"ID":"7c9b2ea9-5417-471b-9e2d-d3e0df537269","Type":"ContainerDied","Data":"08f94091df92461683c05a2c8de45e2b05b17f93381f1875a83d613724300bf2"} Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.313863 4989 scope.go:117] "RemoveContainer" containerID="2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.313527 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ml5rl" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.344380 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.345978 4989 scope.go:117] "RemoveContainer" containerID="31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.351221 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ml5rl"] Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.362132 4989 scope.go:117] "RemoveContainer" containerID="4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.386283 4989 scope.go:117] "RemoveContainer" containerID="2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32" Oct 06 09:35:16 crc kubenswrapper[4989]: E1006 09:35:16.386780 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32\": container with ID starting with 2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32 not found: ID does not exist" containerID="2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.386811 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32"} err="failed to get container status \"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32\": rpc error: code = NotFound desc = could not find container \"2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32\": container with ID starting with 2e3ba2188999665009a0297b7d1b71871f19d13b2389d47e85261caeaab01e32 not found: ID does not exist" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.386835 4989 scope.go:117] "RemoveContainer" containerID="31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31" Oct 06 09:35:16 crc kubenswrapper[4989]: E1006 09:35:16.387044 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31\": container with ID starting with 31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31 not found: ID does not exist" containerID="31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.387068 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31"} err="failed to get container status \"31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31\": rpc error: code = NotFound desc = could not find container \"31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31\": container with ID starting with 31dbda28baadfc9da952a56742a00b280173d78a2392765842dd168bccfd8c31 not found: ID does not exist" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.387081 4989 scope.go:117] "RemoveContainer" containerID="4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919" Oct 06 09:35:16 crc kubenswrapper[4989]: E1006 09:35:16.387289 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919\": container with ID starting with 4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919 not found: ID does not exist" containerID="4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919" Oct 06 09:35:16 crc kubenswrapper[4989]: I1006 09:35:16.387310 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919"} err="failed to get container status \"4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919\": rpc error: code = NotFound desc = could not find container \"4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919\": container with ID starting with 4a9543d1bf86f651f860f9bf557a99377e909af6d7c7e0e949ef88fd5c5ef919 not found: ID does not exist" Oct 06 09:35:17 crc kubenswrapper[4989]: I1006 09:35:17.945003 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" path="/var/lib/kubelet/pods/7c9b2ea9-5417-471b-9e2d-d3e0df537269/volumes" Oct 06 09:35:20 crc kubenswrapper[4989]: I1006 09:35:20.936154 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:35:20 crc kubenswrapper[4989]: E1006 09:35:20.937713 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:35:34 crc kubenswrapper[4989]: I1006 09:35:34.935965 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:35:34 crc kubenswrapper[4989]: E1006 09:35:34.936625 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:35:45 crc kubenswrapper[4989]: I1006 09:35:45.936468 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:35:45 crc kubenswrapper[4989]: E1006 09:35:45.937216 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:35:58 crc kubenswrapper[4989]: I1006 09:35:58.936791 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:35:58 crc kubenswrapper[4989]: E1006 09:35:58.937500 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:36:09 crc kubenswrapper[4989]: I1006 09:36:09.941294 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:36:09 crc kubenswrapper[4989]: E1006 09:36:09.942119 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:36:23 crc kubenswrapper[4989]: I1006 09:36:23.939052 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:36:23 crc kubenswrapper[4989]: E1006 09:36:23.939968 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:36:37 crc kubenswrapper[4989]: I1006 09:36:37.935805 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:36:38 crc kubenswrapper[4989]: I1006 09:36:38.997800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58"} Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.172900 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.173946 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.173966 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.173985 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.173994 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.174010 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="extract-utilities" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174019 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="extract-utilities" Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.174042 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="extract-content" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174051 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="extract-content" Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.174063 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="extract-utilities" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174070 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="extract-utilities" Oct 06 09:38:11 crc kubenswrapper[4989]: E1006 09:38:11.174094 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="extract-content" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174102 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="extract-content" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174350 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="990f7f44-e7e1-4a54-a0ee-ec690b52cca9" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.174385 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c9b2ea9-5417-471b-9e2d-d3e0df537269" containerName="registry-server" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.176002 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.180182 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.327636 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.327718 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2thvh\" (UniqueName: \"kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.327746 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.429209 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.429261 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2thvh\" (UniqueName: \"kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.429281 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.429730 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.429753 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.454889 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2thvh\" (UniqueName: \"kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh\") pod \"redhat-operators-hdjqb\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.495989 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:11 crc kubenswrapper[4989]: I1006 09:38:11.993409 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:12 crc kubenswrapper[4989]: I1006 09:38:12.802722 4989 generic.go:334] "Generic (PLEG): container finished" podID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerID="c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a" exitCode=0 Oct 06 09:38:12 crc kubenswrapper[4989]: I1006 09:38:12.802775 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerDied","Data":"c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a"} Oct 06 09:38:12 crc kubenswrapper[4989]: I1006 09:38:12.803072 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerStarted","Data":"5fc1dc000112f5f57a91c8441a58f38840b3c4d33b96c7f744f8d4044bf43bab"} Oct 06 09:38:14 crc kubenswrapper[4989]: I1006 09:38:14.822988 4989 generic.go:334] "Generic (PLEG): container finished" podID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerID="da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492" exitCode=0 Oct 06 09:38:14 crc kubenswrapper[4989]: I1006 09:38:14.823102 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerDied","Data":"da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492"} Oct 06 09:38:15 crc kubenswrapper[4989]: I1006 09:38:15.835333 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerStarted","Data":"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d"} Oct 06 09:38:15 crc kubenswrapper[4989]: I1006 09:38:15.856117 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hdjqb" podStartSLOduration=2.290733115 podStartE2EDuration="4.8560957s" podCreationTimestamp="2025-10-06 09:38:11 +0000 UTC" firstStartedPulling="2025-10-06 09:38:12.804960136 +0000 UTC m=+3543.594985736" lastFinishedPulling="2025-10-06 09:38:15.370322701 +0000 UTC m=+3546.160348321" observedRunningTime="2025-10-06 09:38:15.855159324 +0000 UTC m=+3546.645184974" watchObservedRunningTime="2025-10-06 09:38:15.8560957 +0000 UTC m=+3546.646121290" Oct 06 09:38:21 crc kubenswrapper[4989]: I1006 09:38:21.496180 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:21 crc kubenswrapper[4989]: I1006 09:38:21.497041 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:21 crc kubenswrapper[4989]: I1006 09:38:21.582467 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:21 crc kubenswrapper[4989]: I1006 09:38:21.957251 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:22 crc kubenswrapper[4989]: I1006 09:38:22.013898 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:23 crc kubenswrapper[4989]: I1006 09:38:23.907915 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hdjqb" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="registry-server" containerID="cri-o://625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d" gracePeriod=2 Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.308030 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.350533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content\") pod \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.350664 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities\") pod \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.350739 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2thvh\" (UniqueName: \"kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh\") pod \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\" (UID: \"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5\") " Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.351617 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities" (OuterVolumeSpecName: "utilities") pod "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" (UID: "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.356875 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh" (OuterVolumeSpecName: "kube-api-access-2thvh") pod "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" (UID: "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5"). InnerVolumeSpecName "kube-api-access-2thvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.453041 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2thvh\" (UniqueName: \"kubernetes.io/projected/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-kube-api-access-2thvh\") on node \"crc\" DevicePath \"\"" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.453084 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.922790 4989 generic.go:334] "Generic (PLEG): container finished" podID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerID="625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d" exitCode=0 Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.923017 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hdjqb" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.923034 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerDied","Data":"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d"} Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.923838 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hdjqb" event={"ID":"08086a64-4d36-4b19-a1d5-7f09fb6f0cc5","Type":"ContainerDied","Data":"5fc1dc000112f5f57a91c8441a58f38840b3c4d33b96c7f744f8d4044bf43bab"} Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.923888 4989 scope.go:117] "RemoveContainer" containerID="625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.943830 4989 scope.go:117] "RemoveContainer" containerID="da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.960966 4989 scope.go:117] "RemoveContainer" containerID="c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.992047 4989 scope.go:117] "RemoveContainer" containerID="625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d" Oct 06 09:38:24 crc kubenswrapper[4989]: E1006 09:38:24.992647 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d\": container with ID starting with 625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d not found: ID does not exist" containerID="625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.992721 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d"} err="failed to get container status \"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d\": rpc error: code = NotFound desc = could not find container \"625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d\": container with ID starting with 625fae426f7e39f202178051f6e57c3a4ebef2051bf5b4684d958cf394d3aa1d not found: ID does not exist" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.992756 4989 scope.go:117] "RemoveContainer" containerID="da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492" Oct 06 09:38:24 crc kubenswrapper[4989]: E1006 09:38:24.993177 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492\": container with ID starting with da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492 not found: ID does not exist" containerID="da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.993217 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492"} err="failed to get container status \"da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492\": rpc error: code = NotFound desc = could not find container \"da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492\": container with ID starting with da2652b32ed2c7be755b19cea06d412a15daf775f393b2bdc231e9690bcc4492 not found: ID does not exist" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.993245 4989 scope.go:117] "RemoveContainer" containerID="c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a" Oct 06 09:38:24 crc kubenswrapper[4989]: E1006 09:38:24.993536 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a\": container with ID starting with c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a not found: ID does not exist" containerID="c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a" Oct 06 09:38:24 crc kubenswrapper[4989]: I1006 09:38:24.993563 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a"} err="failed to get container status \"c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a\": rpc error: code = NotFound desc = could not find container \"c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a\": container with ID starting with c8f5bf4a79b47d5e3dcee001b097d01270b8869148933646d3f13149aaefd20a not found: ID does not exist" Oct 06 09:38:25 crc kubenswrapper[4989]: I1006 09:38:25.286903 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" (UID: "08086a64-4d36-4b19-a1d5-7f09fb6f0cc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:38:25 crc kubenswrapper[4989]: I1006 09:38:25.364732 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:38:25 crc kubenswrapper[4989]: I1006 09:38:25.587639 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:25 crc kubenswrapper[4989]: I1006 09:38:25.601156 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hdjqb"] Oct 06 09:38:25 crc kubenswrapper[4989]: I1006 09:38:25.950758 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" path="/var/lib/kubelet/pods/08086a64-4d36-4b19-a1d5-7f09fb6f0cc5/volumes" Oct 06 09:39:03 crc kubenswrapper[4989]: I1006 09:39:03.934986 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:39:03 crc kubenswrapper[4989]: I1006 09:39:03.938877 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:39:33 crc kubenswrapper[4989]: I1006 09:39:33.935693 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:39:33 crc kubenswrapper[4989]: I1006 09:39:33.936424 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:40:03 crc kubenswrapper[4989]: I1006 09:40:03.935913 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:40:03 crc kubenswrapper[4989]: I1006 09:40:03.936439 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:40:03 crc kubenswrapper[4989]: I1006 09:40:03.948294 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:40:03 crc kubenswrapper[4989]: I1006 09:40:03.949204 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:40:03 crc kubenswrapper[4989]: I1006 09:40:03.949310 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58" gracePeriod=600 Oct 06 09:40:04 crc kubenswrapper[4989]: I1006 09:40:04.770170 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58" exitCode=0 Oct 06 09:40:04 crc kubenswrapper[4989]: I1006 09:40:04.770263 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58"} Oct 06 09:40:04 crc kubenswrapper[4989]: I1006 09:40:04.770489 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c"} Oct 06 09:40:04 crc kubenswrapper[4989]: I1006 09:40:04.770509 4989 scope.go:117] "RemoveContainer" containerID="1fb51f00b61c27724ffcf519e6d3329791507afce8ebce5812e43bd1a029e6f0" Oct 06 09:42:33 crc kubenswrapper[4989]: I1006 09:42:33.935226 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:42:33 crc kubenswrapper[4989]: I1006 09:42:33.935814 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:43:03 crc kubenswrapper[4989]: I1006 09:43:03.935209 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:43:03 crc kubenswrapper[4989]: I1006 09:43:03.935982 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:43:33 crc kubenswrapper[4989]: I1006 09:43:33.935720 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:43:33 crc kubenswrapper[4989]: I1006 09:43:33.936268 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:43:33 crc kubenswrapper[4989]: I1006 09:43:33.949780 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:43:33 crc kubenswrapper[4989]: I1006 09:43:33.950298 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:43:33 crc kubenswrapper[4989]: I1006 09:43:33.950349 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" gracePeriod=600 Oct 06 09:43:34 crc kubenswrapper[4989]: E1006 09:43:34.071561 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:43:34 crc kubenswrapper[4989]: I1006 09:43:34.536900 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" exitCode=0 Oct 06 09:43:34 crc kubenswrapper[4989]: I1006 09:43:34.536942 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c"} Oct 06 09:43:34 crc kubenswrapper[4989]: I1006 09:43:34.536990 4989 scope.go:117] "RemoveContainer" containerID="cf0cfd45daaf105944a66ad5c0927e0f6ec09b298f9cb870ecdc28cabe1ada58" Oct 06 09:43:34 crc kubenswrapper[4989]: I1006 09:43:34.537679 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:43:34 crc kubenswrapper[4989]: E1006 09:43:34.538150 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:43:45 crc kubenswrapper[4989]: I1006 09:43:45.936426 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:43:45 crc kubenswrapper[4989]: E1006 09:43:45.937508 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:44:00 crc kubenswrapper[4989]: I1006 09:44:00.936045 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:44:00 crc kubenswrapper[4989]: E1006 09:44:00.937081 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:44:12 crc kubenswrapper[4989]: I1006 09:44:12.936628 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:44:12 crc kubenswrapper[4989]: E1006 09:44:12.937712 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:44:25 crc kubenswrapper[4989]: I1006 09:44:25.936471 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:44:25 crc kubenswrapper[4989]: E1006 09:44:25.937569 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:44:39 crc kubenswrapper[4989]: I1006 09:44:39.939616 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:44:39 crc kubenswrapper[4989]: E1006 09:44:39.940358 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:44:54 crc kubenswrapper[4989]: I1006 09:44:54.935939 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:44:54 crc kubenswrapper[4989]: E1006 09:44:54.936695 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.166314 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s"] Oct 06 09:45:00 crc kubenswrapper[4989]: E1006 09:45:00.167170 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="extract-utilities" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.167194 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="extract-utilities" Oct 06 09:45:00 crc kubenswrapper[4989]: E1006 09:45:00.167207 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="extract-content" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.167220 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="extract-content" Oct 06 09:45:00 crc kubenswrapper[4989]: E1006 09:45:00.167271 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="registry-server" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.167283 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="registry-server" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.167556 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="08086a64-4d36-4b19-a1d5-7f09fb6f0cc5" containerName="registry-server" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.168408 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.171722 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.172140 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.185384 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s"] Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.320975 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.321223 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.321398 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4n9t\" (UniqueName: \"kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.423278 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.423340 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4n9t\" (UniqueName: \"kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.423390 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.424553 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.432172 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.447176 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4n9t\" (UniqueName: \"kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t\") pod \"collect-profiles-29329065-bv75s\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.495444 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:00 crc kubenswrapper[4989]: I1006 09:45:00.940101 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s"] Oct 06 09:45:01 crc kubenswrapper[4989]: I1006 09:45:01.276910 4989 generic.go:334] "Generic (PLEG): container finished" podID="8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" containerID="2d664d9f2f9a25472a1da0ba5c027a440c085dfbcabfe9de7fde71da65abc7ba" exitCode=0 Oct 06 09:45:01 crc kubenswrapper[4989]: I1006 09:45:01.277009 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" event={"ID":"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a","Type":"ContainerDied","Data":"2d664d9f2f9a25472a1da0ba5c027a440c085dfbcabfe9de7fde71da65abc7ba"} Oct 06 09:45:01 crc kubenswrapper[4989]: I1006 09:45:01.277226 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" event={"ID":"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a","Type":"ContainerStarted","Data":"6978a0b1965d3573bfb5e2e8b8fef6d5ceb0b28888e47ef662f4fa92b8cbbb03"} Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.589878 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.759094 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume\") pod \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.759256 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume\") pod \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.759549 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4n9t\" (UniqueName: \"kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t\") pod \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\" (UID: \"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a\") " Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.760211 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" (UID: "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.761136 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.768206 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t" (OuterVolumeSpecName: "kube-api-access-h4n9t") pod "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" (UID: "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a"). InnerVolumeSpecName "kube-api-access-h4n9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.769502 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" (UID: "8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.862648 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4n9t\" (UniqueName: \"kubernetes.io/projected/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-kube-api-access-h4n9t\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:02 crc kubenswrapper[4989]: I1006 09:45:02.862736 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.297118 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" event={"ID":"8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a","Type":"ContainerDied","Data":"6978a0b1965d3573bfb5e2e8b8fef6d5ceb0b28888e47ef662f4fa92b8cbbb03"} Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.297179 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s" Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.297192 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6978a0b1965d3573bfb5e2e8b8fef6d5ceb0b28888e47ef662f4fa92b8cbbb03" Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.671256 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv"] Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.676331 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329020-dd9dv"] Oct 06 09:45:03 crc kubenswrapper[4989]: I1006 09:45:03.947737 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbc8169-114f-43b8-a6b2-671aa3327536" path="/var/lib/kubelet/pods/7fbc8169-114f-43b8-a6b2-671aa3327536/volumes" Oct 06 09:45:05 crc kubenswrapper[4989]: I1006 09:45:05.936505 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:45:05 crc kubenswrapper[4989]: E1006 09:45:05.937265 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.397323 4989 scope.go:117] "RemoveContainer" containerID="78e06aa2ac711ad850e7a81bd66a5aca2cac653591e37746d8313f6504eadb88" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.700171 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:19 crc kubenswrapper[4989]: E1006 09:45:19.701307 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" containerName="collect-profiles" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.701441 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" containerName="collect-profiles" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.701882 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" containerName="collect-profiles" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.706979 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.713665 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.847332 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb5d8\" (UniqueName: \"kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.847398 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.847422 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.939346 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:45:19 crc kubenswrapper[4989]: E1006 09:45:19.939530 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.949072 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb5d8\" (UniqueName: \"kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.949125 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.949146 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.949597 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.949829 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:19 crc kubenswrapper[4989]: I1006 09:45:19.983205 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb5d8\" (UniqueName: \"kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8\") pod \"community-operators-rn56n\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:20 crc kubenswrapper[4989]: I1006 09:45:20.038177 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:20 crc kubenswrapper[4989]: I1006 09:45:20.504779 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:21 crc kubenswrapper[4989]: I1006 09:45:21.451921 4989 generic.go:334] "Generic (PLEG): container finished" podID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerID="8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672" exitCode=0 Oct 06 09:45:21 crc kubenswrapper[4989]: I1006 09:45:21.452063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerDied","Data":"8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672"} Oct 06 09:45:21 crc kubenswrapper[4989]: I1006 09:45:21.452349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerStarted","Data":"082ae00a6f50125f2ad339fd43b0cfeef6e13452d90fda13bc1abdb5120e924f"} Oct 06 09:45:21 crc kubenswrapper[4989]: I1006 09:45:21.456607 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:45:22 crc kubenswrapper[4989]: I1006 09:45:22.467346 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerStarted","Data":"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db"} Oct 06 09:45:23 crc kubenswrapper[4989]: I1006 09:45:23.478745 4989 generic.go:334] "Generic (PLEG): container finished" podID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerID="f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db" exitCode=0 Oct 06 09:45:23 crc kubenswrapper[4989]: I1006 09:45:23.478812 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerDied","Data":"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db"} Oct 06 09:45:24 crc kubenswrapper[4989]: I1006 09:45:24.490674 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerStarted","Data":"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe"} Oct 06 09:45:24 crc kubenswrapper[4989]: I1006 09:45:24.519417 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rn56n" podStartSLOduration=3.054737072 podStartE2EDuration="5.51939112s" podCreationTimestamp="2025-10-06 09:45:19 +0000 UTC" firstStartedPulling="2025-10-06 09:45:21.456307504 +0000 UTC m=+3972.246333094" lastFinishedPulling="2025-10-06 09:45:23.920961552 +0000 UTC m=+3974.710987142" observedRunningTime="2025-10-06 09:45:24.515393895 +0000 UTC m=+3975.305419475" watchObservedRunningTime="2025-10-06 09:45:24.51939112 +0000 UTC m=+3975.309416710" Oct 06 09:45:30 crc kubenswrapper[4989]: I1006 09:45:30.038978 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:30 crc kubenswrapper[4989]: I1006 09:45:30.039367 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:30 crc kubenswrapper[4989]: I1006 09:45:30.093834 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:30 crc kubenswrapper[4989]: I1006 09:45:30.605828 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:30 crc kubenswrapper[4989]: I1006 09:45:30.652228 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:31 crc kubenswrapper[4989]: I1006 09:45:31.936567 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:45:31 crc kubenswrapper[4989]: E1006 09:45:31.936956 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:45:32 crc kubenswrapper[4989]: I1006 09:45:32.566718 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rn56n" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="registry-server" containerID="cri-o://bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe" gracePeriod=2 Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.053484 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.153212 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content\") pod \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.153359 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities\") pod \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.153441 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb5d8\" (UniqueName: \"kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8\") pod \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\" (UID: \"d37841a3-302e-4f6b-9413-06d0a90e1f1f\") " Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.155064 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities" (OuterVolumeSpecName: "utilities") pod "d37841a3-302e-4f6b-9413-06d0a90e1f1f" (UID: "d37841a3-302e-4f6b-9413-06d0a90e1f1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.160939 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8" (OuterVolumeSpecName: "kube-api-access-jb5d8") pod "d37841a3-302e-4f6b-9413-06d0a90e1f1f" (UID: "d37841a3-302e-4f6b-9413-06d0a90e1f1f"). InnerVolumeSpecName "kube-api-access-jb5d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.203451 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d37841a3-302e-4f6b-9413-06d0a90e1f1f" (UID: "d37841a3-302e-4f6b-9413-06d0a90e1f1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.255056 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.255104 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb5d8\" (UniqueName: \"kubernetes.io/projected/d37841a3-302e-4f6b-9413-06d0a90e1f1f-kube-api-access-jb5d8\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.255123 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37841a3-302e-4f6b-9413-06d0a90e1f1f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.577410 4989 generic.go:334] "Generic (PLEG): container finished" podID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerID="bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe" exitCode=0 Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.577460 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerDied","Data":"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe"} Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.577492 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rn56n" event={"ID":"d37841a3-302e-4f6b-9413-06d0a90e1f1f","Type":"ContainerDied","Data":"082ae00a6f50125f2ad339fd43b0cfeef6e13452d90fda13bc1abdb5120e924f"} Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.577496 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rn56n" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.577509 4989 scope.go:117] "RemoveContainer" containerID="bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.604196 4989 scope.go:117] "RemoveContainer" containerID="f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.622390 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.629929 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rn56n"] Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.695207 4989 scope.go:117] "RemoveContainer" containerID="8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.841055 4989 scope.go:117] "RemoveContainer" containerID="bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe" Oct 06 09:45:33 crc kubenswrapper[4989]: E1006 09:45:33.841809 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe\": container with ID starting with bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe not found: ID does not exist" containerID="bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.841848 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe"} err="failed to get container status \"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe\": rpc error: code = NotFound desc = could not find container \"bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe\": container with ID starting with bc044f82352de1e25f11f082bbfdec16eda770b47789f8decac46f52ed0102fe not found: ID does not exist" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.841882 4989 scope.go:117] "RemoveContainer" containerID="f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db" Oct 06 09:45:33 crc kubenswrapper[4989]: E1006 09:45:33.842217 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db\": container with ID starting with f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db not found: ID does not exist" containerID="f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.842249 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db"} err="failed to get container status \"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db\": rpc error: code = NotFound desc = could not find container \"f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db\": container with ID starting with f0b1e7eb37bac736529807acdf5b53c95e8340c4e1f77980c6e28b903e92a3db not found: ID does not exist" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.842268 4989 scope.go:117] "RemoveContainer" containerID="8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672" Oct 06 09:45:33 crc kubenswrapper[4989]: E1006 09:45:33.842551 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672\": container with ID starting with 8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672 not found: ID does not exist" containerID="8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.842576 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672"} err="failed to get container status \"8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672\": rpc error: code = NotFound desc = could not find container \"8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672\": container with ID starting with 8c0f4e3fb1f16616ecc41062dc25d746246034f88647f3086edad53f3d888672 not found: ID does not exist" Oct 06 09:45:33 crc kubenswrapper[4989]: I1006 09:45:33.949055 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" path="/var/lib/kubelet/pods/d37841a3-302e-4f6b-9413-06d0a90e1f1f/volumes" Oct 06 09:45:42 crc kubenswrapper[4989]: I1006 09:45:42.936201 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:45:42 crc kubenswrapper[4989]: E1006 09:45:42.937234 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:45:53 crc kubenswrapper[4989]: I1006 09:45:53.936512 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:45:53 crc kubenswrapper[4989]: E1006 09:45:53.937488 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:46:06 crc kubenswrapper[4989]: I1006 09:46:06.936315 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:46:06 crc kubenswrapper[4989]: E1006 09:46:06.937443 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:46:19 crc kubenswrapper[4989]: I1006 09:46:19.945690 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:46:19 crc kubenswrapper[4989]: E1006 09:46:19.946889 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:46:34 crc kubenswrapper[4989]: I1006 09:46:34.935725 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:46:34 crc kubenswrapper[4989]: E1006 09:46:34.936436 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:46:46 crc kubenswrapper[4989]: I1006 09:46:46.935764 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:46:46 crc kubenswrapper[4989]: E1006 09:46:46.936862 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:47:00 crc kubenswrapper[4989]: I1006 09:47:00.936116 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:47:00 crc kubenswrapper[4989]: E1006 09:47:00.937290 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:47:13 crc kubenswrapper[4989]: I1006 09:47:13.936273 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:47:13 crc kubenswrapper[4989]: E1006 09:47:13.937193 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:47:28 crc kubenswrapper[4989]: I1006 09:47:28.936518 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:47:28 crc kubenswrapper[4989]: E1006 09:47:28.937541 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:47:42 crc kubenswrapper[4989]: I1006 09:47:42.936014 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:47:42 crc kubenswrapper[4989]: E1006 09:47:42.936929 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:47:57 crc kubenswrapper[4989]: I1006 09:47:57.938150 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:47:57 crc kubenswrapper[4989]: E1006 09:47:57.939197 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:48:12 crc kubenswrapper[4989]: I1006 09:48:12.936243 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:48:12 crc kubenswrapper[4989]: E1006 09:48:12.937108 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:48:23 crc kubenswrapper[4989]: I1006 09:48:23.937397 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:48:23 crc kubenswrapper[4989]: E1006 09:48:23.938194 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:48:37 crc kubenswrapper[4989]: I1006 09:48:37.937338 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:48:38 crc kubenswrapper[4989]: I1006 09:48:38.281949 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3"} Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.280098 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:50:55 crc kubenswrapper[4989]: E1006 09:50:55.281093 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="extract-utilities" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.281114 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="extract-utilities" Oct 06 09:50:55 crc kubenswrapper[4989]: E1006 09:50:55.281139 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="extract-content" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.281150 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="extract-content" Oct 06 09:50:55 crc kubenswrapper[4989]: E1006 09:50:55.281185 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="registry-server" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.281195 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="registry-server" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.281443 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37841a3-302e-4f6b-9413-06d0a90e1f1f" containerName="registry-server" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.283023 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.314028 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.454003 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.454068 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.454185 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpjvx\" (UniqueName: \"kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.555107 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.555185 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.555251 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpjvx\" (UniqueName: \"kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.555716 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.555781 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.575519 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpjvx\" (UniqueName: \"kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx\") pod \"redhat-operators-qprnt\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.623575 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.851736 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.870965 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.872319 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:55 crc kubenswrapper[4989]: I1006 09:50:55.888985 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.062423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm9z6\" (UniqueName: \"kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.062500 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.062734 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.163822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.163889 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.163960 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm9z6\" (UniqueName: \"kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.164366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.164535 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.182704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm9z6\" (UniqueName: \"kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6\") pod \"redhat-marketplace-959m2\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.203387 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.442235 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.503127 4989 generic.go:334] "Generic (PLEG): container finished" podID="926d30dd-8e45-45b5-af66-71a01baac803" containerID="cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908" exitCode=0 Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.503185 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerDied","Data":"cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908"} Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.503210 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerStarted","Data":"c86b4a5d6ea53f5439ed7f94f5fa7d993c6bfeb374ea998ef08445a118fa9f9a"} Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.505542 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:50:56 crc kubenswrapper[4989]: I1006 09:50:56.510816 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerStarted","Data":"f46e9a627eb975849ebf521092f8afe7d237e490671f29a1fb6ba8ae3901617b"} Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.531566 4989 generic.go:334] "Generic (PLEG): container finished" podID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerID="e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2" exitCode=0 Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.531957 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerDied","Data":"e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2"} Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.673500 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.675140 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.688826 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.792039 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.792226 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xfcm\" (UniqueName: \"kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.792330 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.893784 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xfcm\" (UniqueName: \"kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.893858 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.893897 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.894395 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.894462 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:57 crc kubenswrapper[4989]: I1006 09:50:57.917586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xfcm\" (UniqueName: \"kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm\") pod \"certified-operators-gmvxz\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.014250 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.456530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:50:58 crc kubenswrapper[4989]: W1006 09:50:58.492237 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8658d23e_d4c6_41cf_b51e_1799fffa6414.slice/crio-29e2bb4599614ef84943d7ef9dcdb960df755bddf3426a99f8eb9014432ab780 WatchSource:0}: Error finding container 29e2bb4599614ef84943d7ef9dcdb960df755bddf3426a99f8eb9014432ab780: Status 404 returned error can't find the container with id 29e2bb4599614ef84943d7ef9dcdb960df755bddf3426a99f8eb9014432ab780 Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.543221 4989 generic.go:334] "Generic (PLEG): container finished" podID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerID="0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7" exitCode=0 Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.543323 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerDied","Data":"0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7"} Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.545700 4989 generic.go:334] "Generic (PLEG): container finished" podID="926d30dd-8e45-45b5-af66-71a01baac803" containerID="d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586" exitCode=0 Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.545751 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerDied","Data":"d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586"} Oct 06 09:50:58 crc kubenswrapper[4989]: I1006 09:50:58.548950 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerStarted","Data":"29e2bb4599614ef84943d7ef9dcdb960df755bddf3426a99f8eb9014432ab780"} Oct 06 09:50:58 crc kubenswrapper[4989]: E1006 09:50:58.809584 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8658d23e_d4c6_41cf_b51e_1799fffa6414.slice/crio-conmon-b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8658d23e_d4c6_41cf_b51e_1799fffa6414.slice/crio-b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5.scope\": RecentStats: unable to find data in memory cache]" Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.558380 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerStarted","Data":"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13"} Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.560705 4989 generic.go:334] "Generic (PLEG): container finished" podID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerID="b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5" exitCode=0 Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.560768 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerDied","Data":"b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5"} Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.564514 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerStarted","Data":"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d"} Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.579825 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qprnt" podStartSLOduration=2.094368215 podStartE2EDuration="4.57980995s" podCreationTimestamp="2025-10-06 09:50:55 +0000 UTC" firstStartedPulling="2025-10-06 09:50:56.504723908 +0000 UTC m=+4307.294749488" lastFinishedPulling="2025-10-06 09:50:58.990165643 +0000 UTC m=+4309.780191223" observedRunningTime="2025-10-06 09:50:59.576999339 +0000 UTC m=+4310.367024929" watchObservedRunningTime="2025-10-06 09:50:59.57980995 +0000 UTC m=+4310.369835530" Oct 06 09:50:59 crc kubenswrapper[4989]: I1006 09:50:59.601441 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-959m2" podStartSLOduration=3.086465324 podStartE2EDuration="4.601427652s" podCreationTimestamp="2025-10-06 09:50:55 +0000 UTC" firstStartedPulling="2025-10-06 09:50:57.533397949 +0000 UTC m=+4308.323423539" lastFinishedPulling="2025-10-06 09:50:59.048360287 +0000 UTC m=+4309.838385867" observedRunningTime="2025-10-06 09:50:59.597856629 +0000 UTC m=+4310.387882209" watchObservedRunningTime="2025-10-06 09:50:59.601427652 +0000 UTC m=+4310.391453232" Oct 06 09:51:00 crc kubenswrapper[4989]: I1006 09:51:00.577055 4989 generic.go:334] "Generic (PLEG): container finished" podID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerID="a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055" exitCode=0 Oct 06 09:51:00 crc kubenswrapper[4989]: I1006 09:51:00.577132 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerDied","Data":"a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055"} Oct 06 09:51:01 crc kubenswrapper[4989]: I1006 09:51:01.589878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerStarted","Data":"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e"} Oct 06 09:51:01 crc kubenswrapper[4989]: I1006 09:51:01.615607 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gmvxz" podStartSLOduration=3.193166118 podStartE2EDuration="4.615577612s" podCreationTimestamp="2025-10-06 09:50:57 +0000 UTC" firstStartedPulling="2025-10-06 09:50:59.5620825 +0000 UTC m=+4310.352108100" lastFinishedPulling="2025-10-06 09:51:00.984493994 +0000 UTC m=+4311.774519594" observedRunningTime="2025-10-06 09:51:01.610908398 +0000 UTC m=+4312.400933998" watchObservedRunningTime="2025-10-06 09:51:01.615577612 +0000 UTC m=+4312.405603242" Oct 06 09:51:03 crc kubenswrapper[4989]: I1006 09:51:03.935492 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:51:03 crc kubenswrapper[4989]: I1006 09:51:03.937291 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:51:05 crc kubenswrapper[4989]: I1006 09:51:05.624181 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:05 crc kubenswrapper[4989]: I1006 09:51:05.624610 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:05 crc kubenswrapper[4989]: I1006 09:51:05.671178 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:06 crc kubenswrapper[4989]: I1006 09:51:06.203503 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:06 crc kubenswrapper[4989]: I1006 09:51:06.203619 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:06 crc kubenswrapper[4989]: I1006 09:51:06.283029 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:06 crc kubenswrapper[4989]: I1006 09:51:06.688663 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:06 crc kubenswrapper[4989]: I1006 09:51:06.705936 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:08 crc kubenswrapper[4989]: I1006 09:51:08.015287 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:08 crc kubenswrapper[4989]: I1006 09:51:08.015504 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:08 crc kubenswrapper[4989]: I1006 09:51:08.074574 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:51:08 crc kubenswrapper[4989]: I1006 09:51:08.091717 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:08 crc kubenswrapper[4989]: I1006 09:51:08.700332 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.064317 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.064558 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qprnt" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="registry-server" containerID="cri-o://11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13" gracePeriod=2 Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.626118 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668202 4989 generic.go:334] "Generic (PLEG): container finished" podID="926d30dd-8e45-45b5-af66-71a01baac803" containerID="11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13" exitCode=0 Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668269 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerDied","Data":"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13"} Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668337 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qprnt" event={"ID":"926d30dd-8e45-45b5-af66-71a01baac803","Type":"ContainerDied","Data":"c86b4a5d6ea53f5439ed7f94f5fa7d993c6bfeb374ea998ef08445a118fa9f9a"} Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668374 4989 scope.go:117] "RemoveContainer" containerID="11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668825 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qprnt" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.668821 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-959m2" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="registry-server" containerID="cri-o://aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d" gracePeriod=2 Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.695092 4989 scope.go:117] "RemoveContainer" containerID="d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.763229 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content\") pod \"926d30dd-8e45-45b5-af66-71a01baac803\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.763349 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities\") pod \"926d30dd-8e45-45b5-af66-71a01baac803\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.763425 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpjvx\" (UniqueName: \"kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx\") pod \"926d30dd-8e45-45b5-af66-71a01baac803\" (UID: \"926d30dd-8e45-45b5-af66-71a01baac803\") " Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.764787 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities" (OuterVolumeSpecName: "utilities") pod "926d30dd-8e45-45b5-af66-71a01baac803" (UID: "926d30dd-8e45-45b5-af66-71a01baac803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.774238 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx" (OuterVolumeSpecName: "kube-api-access-tpjvx") pod "926d30dd-8e45-45b5-af66-71a01baac803" (UID: "926d30dd-8e45-45b5-af66-71a01baac803"). InnerVolumeSpecName "kube-api-access-tpjvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.777368 4989 scope.go:117] "RemoveContainer" containerID="cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.831814 4989 scope.go:117] "RemoveContainer" containerID="11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13" Oct 06 09:51:09 crc kubenswrapper[4989]: E1006 09:51:09.832407 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13\": container with ID starting with 11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13 not found: ID does not exist" containerID="11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.832476 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13"} err="failed to get container status \"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13\": rpc error: code = NotFound desc = could not find container \"11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13\": container with ID starting with 11be628042a44b56443295e985e1fb806bb67f6bd150cd9242acdf70b71fbb13 not found: ID does not exist" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.832500 4989 scope.go:117] "RemoveContainer" containerID="d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586" Oct 06 09:51:09 crc kubenswrapper[4989]: E1006 09:51:09.833202 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586\": container with ID starting with d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586 not found: ID does not exist" containerID="d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.833491 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586"} err="failed to get container status \"d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586\": rpc error: code = NotFound desc = could not find container \"d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586\": container with ID starting with d9ef8c3599e50229fd9fe9fc49350af2b17a6cfe9feb5349bb6ed8310830e586 not found: ID does not exist" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.833679 4989 scope.go:117] "RemoveContainer" containerID="cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908" Oct 06 09:51:09 crc kubenswrapper[4989]: E1006 09:51:09.834246 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908\": container with ID starting with cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908 not found: ID does not exist" containerID="cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.834280 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908"} err="failed to get container status \"cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908\": rpc error: code = NotFound desc = could not find container \"cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908\": container with ID starting with cb2174be87207ff696b147f7d942ccce4fe64d814077aaf28b311f7952924908 not found: ID does not exist" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.865901 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.866113 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpjvx\" (UniqueName: \"kubernetes.io/projected/926d30dd-8e45-45b5-af66-71a01baac803-kube-api-access-tpjvx\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.885729 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "926d30dd-8e45-45b5-af66-71a01baac803" (UID: "926d30dd-8e45-45b5-af66-71a01baac803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:09 crc kubenswrapper[4989]: I1006 09:51:09.967492 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/926d30dd-8e45-45b5-af66-71a01baac803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.000629 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.006407 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qprnt"] Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.027023 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.169492 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities\") pod \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.169559 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm9z6\" (UniqueName: \"kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6\") pod \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.169860 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content\") pod \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\" (UID: \"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4\") " Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.171044 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities" (OuterVolumeSpecName: "utilities") pod "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" (UID: "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.174265 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6" (OuterVolumeSpecName: "kube-api-access-nm9z6") pod "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" (UID: "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4"). InnerVolumeSpecName "kube-api-access-nm9z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.191129 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" (UID: "0393bdcc-3f70-40ce-a141-e8c6ec95b7b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.271729 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.271780 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.271803 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm9z6\" (UniqueName: \"kubernetes.io/projected/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4-kube-api-access-nm9z6\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.698307 4989 generic.go:334] "Generic (PLEG): container finished" podID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerID="aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d" exitCode=0 Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.698432 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-959m2" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.698453 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerDied","Data":"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d"} Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.698573 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-959m2" event={"ID":"0393bdcc-3f70-40ce-a141-e8c6ec95b7b4","Type":"ContainerDied","Data":"f46e9a627eb975849ebf521092f8afe7d237e490671f29a1fb6ba8ae3901617b"} Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.698608 4989 scope.go:117] "RemoveContainer" containerID="aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.723805 4989 scope.go:117] "RemoveContainer" containerID="0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.758927 4989 scope.go:117] "RemoveContainer" containerID="e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.759408 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.765911 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-959m2"] Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.783611 4989 scope.go:117] "RemoveContainer" containerID="aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d" Oct 06 09:51:10 crc kubenswrapper[4989]: E1006 09:51:10.784032 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d\": container with ID starting with aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d not found: ID does not exist" containerID="aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.784085 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d"} err="failed to get container status \"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d\": rpc error: code = NotFound desc = could not find container \"aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d\": container with ID starting with aa1cce72d3510f472a78c86936f16372608ca5d8ce1ea8477c8ae3f9dab2258d not found: ID does not exist" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.784108 4989 scope.go:117] "RemoveContainer" containerID="0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7" Oct 06 09:51:10 crc kubenswrapper[4989]: E1006 09:51:10.784562 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7\": container with ID starting with 0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7 not found: ID does not exist" containerID="0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.784612 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7"} err="failed to get container status \"0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7\": rpc error: code = NotFound desc = could not find container \"0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7\": container with ID starting with 0b9f986b83a2a3d49a2d28612d1d57e90b4d08eef5d71fd40d19ef62f8e065d7 not found: ID does not exist" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.784639 4989 scope.go:117] "RemoveContainer" containerID="e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2" Oct 06 09:51:10 crc kubenswrapper[4989]: E1006 09:51:10.785057 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2\": container with ID starting with e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2 not found: ID does not exist" containerID="e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2" Oct 06 09:51:10 crc kubenswrapper[4989]: I1006 09:51:10.785143 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2"} err="failed to get container status \"e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2\": rpc error: code = NotFound desc = could not find container \"e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2\": container with ID starting with e73db6ccf5b5b00f2c23b37475976ddf446eeea4a95bbdbccb35b543f5da40a2 not found: ID does not exist" Oct 06 09:51:11 crc kubenswrapper[4989]: I1006 09:51:11.473583 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:51:11 crc kubenswrapper[4989]: I1006 09:51:11.712041 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gmvxz" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="registry-server" containerID="cri-o://5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e" gracePeriod=2 Oct 06 09:51:11 crc kubenswrapper[4989]: I1006 09:51:11.946567 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" path="/var/lib/kubelet/pods/0393bdcc-3f70-40ce-a141-e8c6ec95b7b4/volumes" Oct 06 09:51:11 crc kubenswrapper[4989]: I1006 09:51:11.948149 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="926d30dd-8e45-45b5-af66-71a01baac803" path="/var/lib/kubelet/pods/926d30dd-8e45-45b5-af66-71a01baac803/volumes" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.101140 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.207364 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities\") pod \"8658d23e-d4c6-41cf-b51e-1799fffa6414\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.207471 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xfcm\" (UniqueName: \"kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm\") pod \"8658d23e-d4c6-41cf-b51e-1799fffa6414\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.207519 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content\") pod \"8658d23e-d4c6-41cf-b51e-1799fffa6414\" (UID: \"8658d23e-d4c6-41cf-b51e-1799fffa6414\") " Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.208557 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities" (OuterVolumeSpecName: "utilities") pod "8658d23e-d4c6-41cf-b51e-1799fffa6414" (UID: "8658d23e-d4c6-41cf-b51e-1799fffa6414"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.216212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm" (OuterVolumeSpecName: "kube-api-access-7xfcm") pod "8658d23e-d4c6-41cf-b51e-1799fffa6414" (UID: "8658d23e-d4c6-41cf-b51e-1799fffa6414"). InnerVolumeSpecName "kube-api-access-7xfcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.258304 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8658d23e-d4c6-41cf-b51e-1799fffa6414" (UID: "8658d23e-d4c6-41cf-b51e-1799fffa6414"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.309538 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.309580 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xfcm\" (UniqueName: \"kubernetes.io/projected/8658d23e-d4c6-41cf-b51e-1799fffa6414-kube-api-access-7xfcm\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.309593 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8658d23e-d4c6-41cf-b51e-1799fffa6414-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.730623 4989 generic.go:334] "Generic (PLEG): container finished" podID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerID="5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e" exitCode=0 Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.730700 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerDied","Data":"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e"} Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.730752 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gmvxz" event={"ID":"8658d23e-d4c6-41cf-b51e-1799fffa6414","Type":"ContainerDied","Data":"29e2bb4599614ef84943d7ef9dcdb960df755bddf3426a99f8eb9014432ab780"} Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.730751 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gmvxz" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.730772 4989 scope.go:117] "RemoveContainer" containerID="5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.756713 4989 scope.go:117] "RemoveContainer" containerID="a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.777670 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.778870 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gmvxz"] Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.782109 4989 scope.go:117] "RemoveContainer" containerID="b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.806215 4989 scope.go:117] "RemoveContainer" containerID="5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e" Oct 06 09:51:12 crc kubenswrapper[4989]: E1006 09:51:12.806758 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e\": container with ID starting with 5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e not found: ID does not exist" containerID="5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.806803 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e"} err="failed to get container status \"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e\": rpc error: code = NotFound desc = could not find container \"5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e\": container with ID starting with 5c66d5f3d2b5c35c184b33318d7aa92f5fcf60ac745281ef069a490af50f4d1e not found: ID does not exist" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.806831 4989 scope.go:117] "RemoveContainer" containerID="a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055" Oct 06 09:51:12 crc kubenswrapper[4989]: E1006 09:51:12.807103 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055\": container with ID starting with a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055 not found: ID does not exist" containerID="a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.807132 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055"} err="failed to get container status \"a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055\": rpc error: code = NotFound desc = could not find container \"a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055\": container with ID starting with a86666b70f12bf946afb20c0540c685451710e7e39436903c604e8b4df317055 not found: ID does not exist" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.807157 4989 scope.go:117] "RemoveContainer" containerID="b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5" Oct 06 09:51:12 crc kubenswrapper[4989]: E1006 09:51:12.807430 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5\": container with ID starting with b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5 not found: ID does not exist" containerID="b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5" Oct 06 09:51:12 crc kubenswrapper[4989]: I1006 09:51:12.807466 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5"} err="failed to get container status \"b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5\": rpc error: code = NotFound desc = could not find container \"b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5\": container with ID starting with b9f45c1042d7eb2d3a6ee303463f888a42e56b17939ddd5991db079103b1dfa5 not found: ID does not exist" Oct 06 09:51:13 crc kubenswrapper[4989]: I1006 09:51:13.949518 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" path="/var/lib/kubelet/pods/8658d23e-d4c6-41cf-b51e-1799fffa6414/volumes" Oct 06 09:51:33 crc kubenswrapper[4989]: I1006 09:51:33.935434 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:51:33 crc kubenswrapper[4989]: I1006 09:51:33.936191 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:52:03 crc kubenswrapper[4989]: I1006 09:52:03.935817 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:52:03 crc kubenswrapper[4989]: I1006 09:52:03.936612 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:52:03 crc kubenswrapper[4989]: I1006 09:52:03.948153 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:52:03 crc kubenswrapper[4989]: I1006 09:52:03.948959 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:52:03 crc kubenswrapper[4989]: I1006 09:52:03.949082 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3" gracePeriod=600 Oct 06 09:52:04 crc kubenswrapper[4989]: I1006 09:52:04.163932 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3" exitCode=0 Oct 06 09:52:04 crc kubenswrapper[4989]: I1006 09:52:04.164018 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3"} Oct 06 09:52:04 crc kubenswrapper[4989]: I1006 09:52:04.164269 4989 scope.go:117] "RemoveContainer" containerID="654c2dbdfb053cdf2bd6c9b632e13d6c730f321861a30843c81cead55eb20b0c" Oct 06 09:52:05 crc kubenswrapper[4989]: I1006 09:52:05.188131 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae"} Oct 06 09:54:33 crc kubenswrapper[4989]: I1006 09:54:33.936143 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:54:33 crc kubenswrapper[4989]: I1006 09:54:33.936980 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:55:03 crc kubenswrapper[4989]: I1006 09:55:03.936028 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:55:03 crc kubenswrapper[4989]: I1006 09:55:03.937267 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:55:33 crc kubenswrapper[4989]: I1006 09:55:33.936281 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 09:55:33 crc kubenswrapper[4989]: I1006 09:55:33.937529 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 09:55:33 crc kubenswrapper[4989]: I1006 09:55:33.960850 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 09:55:33 crc kubenswrapper[4989]: I1006 09:55:33.962523 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 09:55:33 crc kubenswrapper[4989]: I1006 09:55:33.962711 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" gracePeriod=600 Oct 06 09:55:34 crc kubenswrapper[4989]: E1006 09:55:34.105685 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:55:34 crc kubenswrapper[4989]: I1006 09:55:34.161353 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" exitCode=0 Oct 06 09:55:34 crc kubenswrapper[4989]: I1006 09:55:34.161441 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae"} Oct 06 09:55:34 crc kubenswrapper[4989]: I1006 09:55:34.161531 4989 scope.go:117] "RemoveContainer" containerID="d9e83db955ba0fb397b608f27bf91aa8b70e9a90ca1f5703a3be7090e68fc7f3" Oct 06 09:55:34 crc kubenswrapper[4989]: I1006 09:55:34.162612 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:55:34 crc kubenswrapper[4989]: E1006 09:55:34.163085 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:55:45 crc kubenswrapper[4989]: I1006 09:55:45.937491 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:55:45 crc kubenswrapper[4989]: E1006 09:55:45.938808 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:55:57 crc kubenswrapper[4989]: I1006 09:55:57.936150 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:55:57 crc kubenswrapper[4989]: E1006 09:55:57.937475 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.059935 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-gfvd9"] Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.067153 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-gfvd9"] Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239095 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-t5hmp"] Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239529 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239557 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239574 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239587 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239609 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239623 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239649 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239687 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239703 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239712 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239731 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239739 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239755 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239764 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239779 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239790 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="extract-content" Oct 06 09:56:05 crc kubenswrapper[4989]: E1006 09:56:05.239813 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.239820 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="extract-utilities" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.240040 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0393bdcc-3f70-40ce-a141-e8c6ec95b7b4" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.240066 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8658d23e-d4c6-41cf-b51e-1799fffa6414" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.240078 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="926d30dd-8e45-45b5-af66-71a01baac803" containerName="registry-server" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.240814 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.244347 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.244388 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.244347 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.245296 4989 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-t6whk" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.272921 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-t5hmp"] Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.326593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.326810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fts85\" (UniqueName: \"kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.326901 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.428068 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fts85\" (UniqueName: \"kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.428202 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.428280 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.428866 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.430091 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.463408 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fts85\" (UniqueName: \"kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85\") pod \"crc-storage-crc-t5hmp\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.572620 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.867139 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-t5hmp"] Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.878810 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 09:56:05 crc kubenswrapper[4989]: I1006 09:56:05.946292 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0326ab5a-9c27-4d96-bd43-971561279956" path="/var/lib/kubelet/pods/0326ab5a-9c27-4d96-bd43-971561279956/volumes" Oct 06 09:56:06 crc kubenswrapper[4989]: I1006 09:56:06.510002 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5hmp" event={"ID":"f0362f22-6da6-4a35-9a07-dc3921520d95","Type":"ContainerStarted","Data":"fafb67d8d67febc68341e64a973d54ecddeb76eb9e15813cbef04d0fba2d518d"} Oct 06 09:56:07 crc kubenswrapper[4989]: I1006 09:56:07.524693 4989 generic.go:334] "Generic (PLEG): container finished" podID="f0362f22-6da6-4a35-9a07-dc3921520d95" containerID="e46f6268a8f0cd083838f3377bb458f404b3dc276905773e00620ec1986b6b4a" exitCode=0 Oct 06 09:56:07 crc kubenswrapper[4989]: I1006 09:56:07.524843 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5hmp" event={"ID":"f0362f22-6da6-4a35-9a07-dc3921520d95","Type":"ContainerDied","Data":"e46f6268a8f0cd083838f3377bb458f404b3dc276905773e00620ec1986b6b4a"} Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.882209 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.988247 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage\") pod \"f0362f22-6da6-4a35-9a07-dc3921520d95\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.988434 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fts85\" (UniqueName: \"kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85\") pod \"f0362f22-6da6-4a35-9a07-dc3921520d95\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.988550 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt\") pod \"f0362f22-6da6-4a35-9a07-dc3921520d95\" (UID: \"f0362f22-6da6-4a35-9a07-dc3921520d95\") " Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.988980 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "f0362f22-6da6-4a35-9a07-dc3921520d95" (UID: "f0362f22-6da6-4a35-9a07-dc3921520d95"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:56:08 crc kubenswrapper[4989]: I1006 09:56:08.995942 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85" (OuterVolumeSpecName: "kube-api-access-fts85") pod "f0362f22-6da6-4a35-9a07-dc3921520d95" (UID: "f0362f22-6da6-4a35-9a07-dc3921520d95"). InnerVolumeSpecName "kube-api-access-fts85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.028073 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "f0362f22-6da6-4a35-9a07-dc3921520d95" (UID: "f0362f22-6da6-4a35-9a07-dc3921520d95"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.089935 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fts85\" (UniqueName: \"kubernetes.io/projected/f0362f22-6da6-4a35-9a07-dc3921520d95-kube-api-access-fts85\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.089974 4989 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f0362f22-6da6-4a35-9a07-dc3921520d95-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.089986 4989 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f0362f22-6da6-4a35-9a07-dc3921520d95-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.567239 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-t5hmp" event={"ID":"f0362f22-6da6-4a35-9a07-dc3921520d95","Type":"ContainerDied","Data":"fafb67d8d67febc68341e64a973d54ecddeb76eb9e15813cbef04d0fba2d518d"} Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.567299 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fafb67d8d67febc68341e64a973d54ecddeb76eb9e15813cbef04d0fba2d518d" Oct 06 09:56:09 crc kubenswrapper[4989]: I1006 09:56:09.567330 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-t5hmp" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.248648 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:10 crc kubenswrapper[4989]: E1006 09:56:10.249038 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0362f22-6da6-4a35-9a07-dc3921520d95" containerName="storage" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.249052 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0362f22-6da6-4a35-9a07-dc3921520d95" containerName="storage" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.249200 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0362f22-6da6-4a35-9a07-dc3921520d95" containerName="storage" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.250318 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.265767 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.315107 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.315523 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.315580 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppkgh\" (UniqueName: \"kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.417195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppkgh\" (UniqueName: \"kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.417350 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.417463 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.417939 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.418169 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.483647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppkgh\" (UniqueName: \"kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh\") pod \"community-operators-g96mn\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:10 crc kubenswrapper[4989]: I1006 09:56:10.606926 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.187344 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:11 crc kubenswrapper[4989]: W1006 09:56:11.190939 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86bdbb26_57e8_49ba_8379_ee9e354c71ad.slice/crio-eeeafec88e5d9ea836390ee61738d1dfdeeecdced0ae4d81ee4792cb4d48ec36 WatchSource:0}: Error finding container eeeafec88e5d9ea836390ee61738d1dfdeeecdced0ae4d81ee4792cb4d48ec36: Status 404 returned error can't find the container with id eeeafec88e5d9ea836390ee61738d1dfdeeecdced0ae4d81ee4792cb4d48ec36 Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.455302 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-t5hmp"] Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.461376 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-t5hmp"] Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.588771 4989 generic.go:334] "Generic (PLEG): container finished" podID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerID="8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e" exitCode=0 Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.588838 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerDied","Data":"8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e"} Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.588884 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerStarted","Data":"eeeafec88e5d9ea836390ee61738d1dfdeeecdced0ae4d81ee4792cb4d48ec36"} Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.657634 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-2qzbc"] Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.661120 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.668038 4989 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-t6whk" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.668313 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.668497 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.668785 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.669976 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2qzbc"] Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.745938 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.746234 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.746452 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwfpx\" (UniqueName: \"kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.847937 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.848181 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.848285 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwfpx\" (UniqueName: \"kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.848516 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.850008 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.881761 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwfpx\" (UniqueName: \"kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx\") pod \"crc-storage-crc-2qzbc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.959263 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0362f22-6da6-4a35-9a07-dc3921520d95" path="/var/lib/kubelet/pods/f0362f22-6da6-4a35-9a07-dc3921520d95/volumes" Oct 06 09:56:11 crc kubenswrapper[4989]: I1006 09:56:11.994323 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:12 crc kubenswrapper[4989]: I1006 09:56:12.888816 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-2qzbc"] Oct 06 09:56:12 crc kubenswrapper[4989]: I1006 09:56:12.936346 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:56:12 crc kubenswrapper[4989]: E1006 09:56:12.936744 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:56:13 crc kubenswrapper[4989]: I1006 09:56:13.611582 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2qzbc" event={"ID":"66e35298-ec05-4189-bb9d-593077bf8ebc","Type":"ContainerStarted","Data":"d70c80542688dc0f2a667b14b649f98072e559fe528b9a46af35755222ecac9a"} Oct 06 09:56:14 crc kubenswrapper[4989]: I1006 09:56:14.625355 4989 generic.go:334] "Generic (PLEG): container finished" podID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerID="1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6" exitCode=0 Oct 06 09:56:14 crc kubenswrapper[4989]: I1006 09:56:14.625500 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerDied","Data":"1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6"} Oct 06 09:56:14 crc kubenswrapper[4989]: I1006 09:56:14.629462 4989 generic.go:334] "Generic (PLEG): container finished" podID="66e35298-ec05-4189-bb9d-593077bf8ebc" containerID="d09f4590fe484a94c5f2fe4885351bf655e3d3ee23b0b941d1d2d05fdfc9237b" exitCode=0 Oct 06 09:56:14 crc kubenswrapper[4989]: I1006 09:56:14.629530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2qzbc" event={"ID":"66e35298-ec05-4189-bb9d-593077bf8ebc","Type":"ContainerDied","Data":"d09f4590fe484a94c5f2fe4885351bf655e3d3ee23b0b941d1d2d05fdfc9237b"} Oct 06 09:56:15 crc kubenswrapper[4989]: I1006 09:56:15.640519 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerStarted","Data":"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe"} Oct 06 09:56:15 crc kubenswrapper[4989]: I1006 09:56:15.674362 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g96mn" podStartSLOduration=2.215213801 podStartE2EDuration="5.674341959s" podCreationTimestamp="2025-10-06 09:56:10 +0000 UTC" firstStartedPulling="2025-10-06 09:56:11.591158102 +0000 UTC m=+4622.381183682" lastFinishedPulling="2025-10-06 09:56:15.05028623 +0000 UTC m=+4625.840311840" observedRunningTime="2025-10-06 09:56:15.669207361 +0000 UTC m=+4626.459232981" watchObservedRunningTime="2025-10-06 09:56:15.674341959 +0000 UTC m=+4626.464367549" Oct 06 09:56:15 crc kubenswrapper[4989]: I1006 09:56:15.955030 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.030665 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwfpx\" (UniqueName: \"kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx\") pod \"66e35298-ec05-4189-bb9d-593077bf8ebc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.030715 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt\") pod \"66e35298-ec05-4189-bb9d-593077bf8ebc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.030749 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage\") pod \"66e35298-ec05-4189-bb9d-593077bf8ebc\" (UID: \"66e35298-ec05-4189-bb9d-593077bf8ebc\") " Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.030857 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "66e35298-ec05-4189-bb9d-593077bf8ebc" (UID: "66e35298-ec05-4189-bb9d-593077bf8ebc"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.031194 4989 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/66e35298-ec05-4189-bb9d-593077bf8ebc-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.037734 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx" (OuterVolumeSpecName: "kube-api-access-bwfpx") pod "66e35298-ec05-4189-bb9d-593077bf8ebc" (UID: "66e35298-ec05-4189-bb9d-593077bf8ebc"). InnerVolumeSpecName "kube-api-access-bwfpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.049012 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "66e35298-ec05-4189-bb9d-593077bf8ebc" (UID: "66e35298-ec05-4189-bb9d-593077bf8ebc"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.132859 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwfpx\" (UniqueName: \"kubernetes.io/projected/66e35298-ec05-4189-bb9d-593077bf8ebc-kube-api-access-bwfpx\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.132898 4989 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/66e35298-ec05-4189-bb9d-593077bf8ebc-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.651020 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-2qzbc" Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.651012 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-2qzbc" event={"ID":"66e35298-ec05-4189-bb9d-593077bf8ebc","Type":"ContainerDied","Data":"d70c80542688dc0f2a667b14b649f98072e559fe528b9a46af35755222ecac9a"} Oct 06 09:56:16 crc kubenswrapper[4989]: I1006 09:56:16.652459 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d70c80542688dc0f2a667b14b649f98072e559fe528b9a46af35755222ecac9a" Oct 06 09:56:19 crc kubenswrapper[4989]: I1006 09:56:19.654198 4989 scope.go:117] "RemoveContainer" containerID="45967ea767aea12bb58304edab0ce67ec5c5ad5dfd188e799d07704e13309789" Oct 06 09:56:20 crc kubenswrapper[4989]: I1006 09:56:20.607328 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:20 crc kubenswrapper[4989]: I1006 09:56:20.607916 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:20 crc kubenswrapper[4989]: I1006 09:56:20.676082 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:20 crc kubenswrapper[4989]: I1006 09:56:20.768622 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:20 crc kubenswrapper[4989]: I1006 09:56:20.924935 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:22 crc kubenswrapper[4989]: I1006 09:56:22.721784 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g96mn" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="registry-server" containerID="cri-o://f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe" gracePeriod=2 Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.254271 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.353242 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppkgh\" (UniqueName: \"kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh\") pod \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.353363 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities\") pod \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.353408 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content\") pod \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\" (UID: \"86bdbb26-57e8-49ba-8379-ee9e354c71ad\") " Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.354775 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities" (OuterVolumeSpecName: "utilities") pod "86bdbb26-57e8-49ba-8379-ee9e354c71ad" (UID: "86bdbb26-57e8-49ba-8379-ee9e354c71ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.361603 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh" (OuterVolumeSpecName: "kube-api-access-ppkgh") pod "86bdbb26-57e8-49ba-8379-ee9e354c71ad" (UID: "86bdbb26-57e8-49ba-8379-ee9e354c71ad"). InnerVolumeSpecName "kube-api-access-ppkgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.406999 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86bdbb26-57e8-49ba-8379-ee9e354c71ad" (UID: "86bdbb26-57e8-49ba-8379-ee9e354c71ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.455005 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.455039 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppkgh\" (UniqueName: \"kubernetes.io/projected/86bdbb26-57e8-49ba-8379-ee9e354c71ad-kube-api-access-ppkgh\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.455057 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86bdbb26-57e8-49ba-8379-ee9e354c71ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.742073 4989 generic.go:334] "Generic (PLEG): container finished" podID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerID="f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe" exitCode=0 Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.742186 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g96mn" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.742220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerDied","Data":"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe"} Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.742768 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g96mn" event={"ID":"86bdbb26-57e8-49ba-8379-ee9e354c71ad","Type":"ContainerDied","Data":"eeeafec88e5d9ea836390ee61738d1dfdeeecdced0ae4d81ee4792cb4d48ec36"} Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.742805 4989 scope.go:117] "RemoveContainer" containerID="f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.794452 4989 scope.go:117] "RemoveContainer" containerID="1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6" Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.810800 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.817017 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g96mn"] Oct 06 09:56:23 crc kubenswrapper[4989]: I1006 09:56:23.955333 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" path="/var/lib/kubelet/pods/86bdbb26-57e8-49ba-8379-ee9e354c71ad/volumes" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.199998 4989 scope.go:117] "RemoveContainer" containerID="8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.252366 4989 scope.go:117] "RemoveContainer" containerID="f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe" Oct 06 09:56:24 crc kubenswrapper[4989]: E1006 09:56:24.253573 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe\": container with ID starting with f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe not found: ID does not exist" containerID="f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.253715 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe"} err="failed to get container status \"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe\": rpc error: code = NotFound desc = could not find container \"f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe\": container with ID starting with f3cea13c4e1bcbdb8910738269314310fbdedd5ced571467ea46f999f303fcbe not found: ID does not exist" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.253876 4989 scope.go:117] "RemoveContainer" containerID="1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6" Oct 06 09:56:24 crc kubenswrapper[4989]: E1006 09:56:24.254874 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6\": container with ID starting with 1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6 not found: ID does not exist" containerID="1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.254941 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6"} err="failed to get container status \"1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6\": rpc error: code = NotFound desc = could not find container \"1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6\": container with ID starting with 1a0dd39bbfaf6e7414e11612724dcd59704c717df5381d6fcd36982ca5708ca6 not found: ID does not exist" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.254986 4989 scope.go:117] "RemoveContainer" containerID="8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e" Oct 06 09:56:24 crc kubenswrapper[4989]: E1006 09:56:24.255789 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e\": container with ID starting with 8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e not found: ID does not exist" containerID="8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e" Oct 06 09:56:24 crc kubenswrapper[4989]: I1006 09:56:24.255867 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e"} err="failed to get container status \"8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e\": rpc error: code = NotFound desc = could not find container \"8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e\": container with ID starting with 8de8a933dc45fec7b0aaec000e154914ebae9ac5d703f06084ef77a9404b689e not found: ID does not exist" Oct 06 09:56:27 crc kubenswrapper[4989]: I1006 09:56:27.936244 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:56:27 crc kubenswrapper[4989]: E1006 09:56:27.937093 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:56:39 crc kubenswrapper[4989]: I1006 09:56:39.947411 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:56:39 crc kubenswrapper[4989]: E1006 09:56:39.948750 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:56:52 crc kubenswrapper[4989]: I1006 09:56:52.936631 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:56:52 crc kubenswrapper[4989]: E1006 09:56:52.937799 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:57:05 crc kubenswrapper[4989]: I1006 09:57:05.935982 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:57:05 crc kubenswrapper[4989]: E1006 09:57:05.937281 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:57:16 crc kubenswrapper[4989]: I1006 09:57:16.936863 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:57:16 crc kubenswrapper[4989]: E1006 09:57:16.938201 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:57:28 crc kubenswrapper[4989]: I1006 09:57:28.936842 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:57:28 crc kubenswrapper[4989]: E1006 09:57:28.937924 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:57:41 crc kubenswrapper[4989]: I1006 09:57:41.936766 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:57:41 crc kubenswrapper[4989]: E1006 09:57:41.938018 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:57:54 crc kubenswrapper[4989]: I1006 09:57:54.936324 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:57:54 crc kubenswrapper[4989]: E1006 09:57:54.937518 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:58:05 crc kubenswrapper[4989]: I1006 09:58:05.942408 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:58:05 crc kubenswrapper[4989]: E1006 09:58:05.943517 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:58:16 crc kubenswrapper[4989]: I1006 09:58:16.936855 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:58:16 crc kubenswrapper[4989]: E1006 09:58:16.938078 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:58:27 crc kubenswrapper[4989]: I1006 09:58:27.936601 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:58:27 crc kubenswrapper[4989]: E1006 09:58:27.937786 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:58:42 crc kubenswrapper[4989]: I1006 09:58:42.937597 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:58:42 crc kubenswrapper[4989]: E1006 09:58:42.939063 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:58:53 crc kubenswrapper[4989]: I1006 09:58:53.937301 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:58:53 crc kubenswrapper[4989]: E1006 09:58:53.938767 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:59:06 crc kubenswrapper[4989]: I1006 09:59:06.936583 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:59:06 crc kubenswrapper[4989]: E1006 09:59:06.937847 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:59:20 crc kubenswrapper[4989]: I1006 09:59:20.935779 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:59:20 crc kubenswrapper[4989]: E1006 09:59:20.936630 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.874920 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:23 crc kubenswrapper[4989]: E1006 09:59:23.875638 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="registry-server" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875671 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="registry-server" Oct 06 09:59:23 crc kubenswrapper[4989]: E1006 09:59:23.875686 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="extract-utilities" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875693 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="extract-utilities" Oct 06 09:59:23 crc kubenswrapper[4989]: E1006 09:59:23.875701 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e35298-ec05-4189-bb9d-593077bf8ebc" containerName="storage" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875706 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e35298-ec05-4189-bb9d-593077bf8ebc" containerName="storage" Oct 06 09:59:23 crc kubenswrapper[4989]: E1006 09:59:23.875735 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="extract-content" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875741 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="extract-content" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875862 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="86bdbb26-57e8-49ba-8379-ee9e354c71ad" containerName="registry-server" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.875883 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e35298-ec05-4189-bb9d-593077bf8ebc" containerName="storage" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.876574 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.879312 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.879377 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-tdl8n" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.879374 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.879323 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.886315 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 06 09:59:23 crc kubenswrapper[4989]: I1006 09:59:23.888549 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.008483 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.008560 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t82t7\" (UniqueName: \"kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.009089 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.035092 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.036570 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.056832 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.110053 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.110154 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.110179 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t82t7\" (UniqueName: \"kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.111915 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.112477 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.134912 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t82t7\" (UniqueName: \"kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7\") pod \"dnsmasq-dns-5d7b5456f5-zkqcl\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.200735 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.211331 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.211393 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bktln\" (UniqueName: \"kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.211610 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.312955 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.313017 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bktln\" (UniqueName: \"kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.313050 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.313916 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.314428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.338392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bktln\" (UniqueName: \"kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln\") pod \"dnsmasq-dns-98ddfc8f-d5jbr\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.357709 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.686752 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.855139 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.856159 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.858444 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-k5nfm" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.858472 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.897101 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.909145 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.948947 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.951872 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.953495 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x4jfh" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.955569 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.955734 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.955853 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.955897 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 06 09:59:24 crc kubenswrapper[4989]: I1006 09:59:24.982229 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.023429 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-kolla-config\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.023467 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-config-data\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.023491 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tzdw\" (UniqueName: \"kubernetes.io/projected/ea1fbbbe-057b-4682-8900-d5c17675a000-kube-api-access-7tzdw\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.124970 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125033 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-kolla-config\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125059 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-config-data\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125078 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tzdw\" (UniqueName: \"kubernetes.io/projected/ea1fbbbe-057b-4682-8900-d5c17675a000-kube-api-access-7tzdw\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125142 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125195 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125219 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125240 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125294 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxhwj\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.125313 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.126861 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-kolla-config\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.127773 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea1fbbbe-057b-4682-8900-d5c17675a000-config-data\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.149619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tzdw\" (UniqueName: \"kubernetes.io/projected/ea1fbbbe-057b-4682-8900-d5c17675a000-kube-api-access-7tzdw\") pod \"memcached-0\" (UID: \"ea1fbbbe-057b-4682-8900-d5c17675a000\") " pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.185689 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.211398 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.212530 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.214633 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hx85m" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.214822 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.215116 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.215283 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.215402 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226790 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226861 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxhwj\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226885 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226909 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226942 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226961 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.226994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.227013 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.227032 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.230077 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.230496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.230621 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.230787 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.231315 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.234473 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.246752 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.251434 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.254221 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.254674 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0927e9832013118ef8b604f038c0b9a415a2e415e922403e4d496bf8d6311ab9/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.258000 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxhwj\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.283461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.292308 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328385 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328479 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328555 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328627 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328729 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328777 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328835 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.328901 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwsmf\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.430688 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.430747 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.430773 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.430829 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.431382 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.432776 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.432810 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwsmf\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.432839 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.432879 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.432920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.434346 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.434598 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.435274 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.438318 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.438359 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/21bc325a0270dfafae871c98be722654f90969b8795d181372052a53810d41e3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.442772 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.444876 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.445045 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.447202 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.449217 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwsmf\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: W1006 09:59:25.461010 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea1fbbbe_057b_4682_8900_d5c17675a000.slice/crio-0e0199a7329ff181724d3d84cef6699660ad9d2920ff5239bfcc713df77ee0cd WatchSource:0}: Error finding container 0e0199a7329ff181724d3d84cef6699660ad9d2920ff5239bfcc713df77ee0cd: Status 404 returned error can't find the container with id 0e0199a7329ff181724d3d84cef6699660ad9d2920ff5239bfcc713df77ee0cd Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.476124 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.495496 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ea1fbbbe-057b-4682-8900-d5c17675a000","Type":"ContainerStarted","Data":"0e0199a7329ff181724d3d84cef6699660ad9d2920ff5239bfcc713df77ee0cd"} Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.500359 4989 generic.go:334] "Generic (PLEG): container finished" podID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerID="4f9b4b3a4072da73ea2b0a384dbf5eec6fb4d729c64c721b2d7d2af75772f274" exitCode=0 Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.500432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" event={"ID":"480bbde2-3ca3-4e6f-bc34-e530774cb253","Type":"ContainerDied","Data":"4f9b4b3a4072da73ea2b0a384dbf5eec6fb4d729c64c721b2d7d2af75772f274"} Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.500452 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" event={"ID":"480bbde2-3ca3-4e6f-bc34-e530774cb253","Type":"ContainerStarted","Data":"af181ac88906bda571613a6a3f23400ff95d84d0516de748928d6c0fece47bb1"} Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.506478 4989 generic.go:334] "Generic (PLEG): container finished" podID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerID="ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4" exitCode=0 Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.506670 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" event={"ID":"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7","Type":"ContainerDied","Data":"ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4"} Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.506793 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" event={"ID":"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7","Type":"ContainerStarted","Data":"065a12502ced34bdb19e84f7c3048e4254ae14fa6f70bd31042668e14e97d49c"} Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.649593 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.649927 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.652523 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.657470 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.657607 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.657688 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jqw6p" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.660087 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.660128 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.662107 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.665040 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739544 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739691 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739724 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739771 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739788 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739846 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.739891 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.740515 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-468d7\" (UniqueName: \"kubernetes.io/projected/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kube-api-access-468d7\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.740544 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.763947 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: W1006 09:59:25.776303 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod314237cb_dbf0_46d7_9830_6c1c10736eb8.slice/crio-03d87d27a9096d18ed0c265b66eb3283360f5b51cdbd85ec98d99680a1a08d62 WatchSource:0}: Error finding container 03d87d27a9096d18ed0c265b66eb3283360f5b51cdbd85ec98d99680a1a08d62: Status 404 returned error can't find the container with id 03d87d27a9096d18ed0c265b66eb3283360f5b51cdbd85ec98d99680a1a08d62 Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.843554 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844201 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844348 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-468d7\" (UniqueName: \"kubernetes.io/projected/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kube-api-access-468d7\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844385 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844518 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844721 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.844926 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.845360 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.845612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.848304 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.848603 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.848702 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/057f129a46a9d1735158619ce6a9142bcec9f59076727fed830c4eb3af56be43/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.849355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.851810 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.857985 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.860232 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.860511 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-468d7\" (UniqueName: \"kubernetes.io/projected/6b4c9853-f4c0-4e59-9d5a-78af6bf457b1-kube-api-access-468d7\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.881345 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.884460 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a785dca9-88ae-4900-8ba1-3fc76889d430\") pod \"openstack-cell1-galera-0\" (UID: \"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1\") " pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:25 crc kubenswrapper[4989]: W1006 09:59:25.889797 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaaf8776_5729_4000_8e9d_153c983587cb.slice/crio-15c00c981afabd675a6e33f0d65e1f75ebc04db345659242659f5c27012fa07c WatchSource:0}: Error finding container 15c00c981afabd675a6e33f0d65e1f75ebc04db345659242659f5c27012fa07c: Status 404 returned error can't find the container with id 15c00c981afabd675a6e33f0d65e1f75ebc04db345659242659f5c27012fa07c Oct 06 09:59:25 crc kubenswrapper[4989]: I1006 09:59:25.975990 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.054585 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.057428 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.059509 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.059791 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4m2hj" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.062435 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.067046 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.067630 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.154794 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-config-data-default\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.154864 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.154908 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fd317fc0-8e84-4a97-94d7-eff210545286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd317fc0-8e84-4a97-94d7-eff210545286\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.154955 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-kolla-config\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.154988 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.155161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-secrets\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.155462 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/678e9748-bfb5-47c1-8788-083b5189145c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.155531 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt5r2\" (UniqueName: \"kubernetes.io/projected/678e9748-bfb5-47c1-8788-083b5189145c-kube-api-access-zt5r2\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.155677 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259124 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-secrets\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259548 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/678e9748-bfb5-47c1-8788-083b5189145c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259583 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt5r2\" (UniqueName: \"kubernetes.io/projected/678e9748-bfb5-47c1-8788-083b5189145c-kube-api-access-zt5r2\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259615 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259683 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-config-data-default\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259709 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259743 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fd317fc0-8e84-4a97-94d7-eff210545286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd317fc0-8e84-4a97-94d7-eff210545286\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259768 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-kolla-config\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.259793 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.260482 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/678e9748-bfb5-47c1-8788-083b5189145c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.261544 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-kolla-config\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.261992 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-config-data-default\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.262722 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/678e9748-bfb5-47c1-8788-083b5189145c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.263969 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.264026 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fd317fc0-8e84-4a97-94d7-eff210545286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd317fc0-8e84-4a97-94d7-eff210545286\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/89e9caff956d68ea2b556569e12b42ca2f6d2676020faa35e175f579b4b28788/globalmount\"" pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.267240 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-secrets\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.269860 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.275040 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678e9748-bfb5-47c1-8788-083b5189145c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.283397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt5r2\" (UniqueName: \"kubernetes.io/projected/678e9748-bfb5-47c1-8788-083b5189145c-kube-api-access-zt5r2\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.312289 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fd317fc0-8e84-4a97-94d7-eff210545286\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd317fc0-8e84-4a97-94d7-eff210545286\") pod \"openstack-galera-0\" (UID: \"678e9748-bfb5-47c1-8788-083b5189145c\") " pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.377248 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.472722 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.520073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" event={"ID":"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7","Type":"ContainerStarted","Data":"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.520217 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.522800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerStarted","Data":"15c00c981afabd675a6e33f0d65e1f75ebc04db345659242659f5c27012fa07c"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.524569 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"ea1fbbbe-057b-4682-8900-d5c17675a000","Type":"ContainerStarted","Data":"266082a294b7d5273dfb66e9cc4bb3d11ac1d6471ca78674140bef8c772530f0"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.524702 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.531241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerStarted","Data":"03d87d27a9096d18ed0c265b66eb3283360f5b51cdbd85ec98d99680a1a08d62"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.532272 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1","Type":"ContainerStarted","Data":"1f2b26bcc06b83ae4489b6e94821a2882f3791c0e7ea617e713f1473344d964b"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.535544 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" event={"ID":"480bbde2-3ca3-4e6f-bc34-e530774cb253","Type":"ContainerStarted","Data":"19844a6861d6f0a7afd3a0081319ef88bed9fc7e2b76b300c54e7422ea204681"} Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.535845 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.566549 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" podStartSLOduration=2.566528763 podStartE2EDuration="2.566528763s" podCreationTimestamp="2025-10-06 09:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:26.561288482 +0000 UTC m=+4817.351314062" watchObservedRunningTime="2025-10-06 09:59:26.566528763 +0000 UTC m=+4817.356554343" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.566750 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" podStartSLOduration=3.566746629 podStartE2EDuration="3.566746629s" podCreationTimestamp="2025-10-06 09:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:26.543398876 +0000 UTC m=+4817.333424466" watchObservedRunningTime="2025-10-06 09:59:26.566746629 +0000 UTC m=+4817.356772209" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.582579 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.582558305 podStartE2EDuration="2.582558305s" podCreationTimestamp="2025-10-06 09:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:26.579591679 +0000 UTC m=+4817.369617259" watchObservedRunningTime="2025-10-06 09:59:26.582558305 +0000 UTC m=+4817.372583895" Oct 06 09:59:26 crc kubenswrapper[4989]: I1006 09:59:26.635268 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 06 09:59:26 crc kubenswrapper[4989]: W1006 09:59:26.640733 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod678e9748_bfb5_47c1_8788_083b5189145c.slice/crio-0cdf41062cd868f25118536cadd5eede96e10db4b52ea6fef5fa33b0d8f049fe WatchSource:0}: Error finding container 0cdf41062cd868f25118536cadd5eede96e10db4b52ea6fef5fa33b0d8f049fe: Status 404 returned error can't find the container with id 0cdf41062cd868f25118536cadd5eede96e10db4b52ea6fef5fa33b0d8f049fe Oct 06 09:59:27 crc kubenswrapper[4989]: I1006 09:59:27.544443 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1","Type":"ContainerStarted","Data":"cbe64e4434544611a3dd5ae2d7a07708c915dff86ae4f4dc2633e94eb3fa9e8a"} Oct 06 09:59:27 crc kubenswrapper[4989]: I1006 09:59:27.546160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"678e9748-bfb5-47c1-8788-083b5189145c","Type":"ContainerStarted","Data":"0cdf41062cd868f25118536cadd5eede96e10db4b52ea6fef5fa33b0d8f049fe"} Oct 06 09:59:28 crc kubenswrapper[4989]: I1006 09:59:28.557088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerStarted","Data":"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b"} Oct 06 09:59:28 crc kubenswrapper[4989]: I1006 09:59:28.558443 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerStarted","Data":"cf1cf42ad2c6da61ee8f7bb5a6ecda8e90f92656686681e24f2275818891c885"} Oct 06 09:59:28 crc kubenswrapper[4989]: I1006 09:59:28.561256 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"678e9748-bfb5-47c1-8788-083b5189145c","Type":"ContainerStarted","Data":"d5fbfb18bc8dae146b5cbcdb5bea6e9020c81a674b5a478f89e75e8bf06f84ee"} Oct 06 09:59:30 crc kubenswrapper[4989]: I1006 09:59:30.187947 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 06 09:59:30 crc kubenswrapper[4989]: I1006 09:59:30.586137 4989 generic.go:334] "Generic (PLEG): container finished" podID="6b4c9853-f4c0-4e59-9d5a-78af6bf457b1" containerID="cbe64e4434544611a3dd5ae2d7a07708c915dff86ae4f4dc2633e94eb3fa9e8a" exitCode=0 Oct 06 09:59:30 crc kubenswrapper[4989]: I1006 09:59:30.586269 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1","Type":"ContainerDied","Data":"cbe64e4434544611a3dd5ae2d7a07708c915dff86ae4f4dc2633e94eb3fa9e8a"} Oct 06 09:59:31 crc kubenswrapper[4989]: I1006 09:59:31.596929 4989 generic.go:334] "Generic (PLEG): container finished" podID="678e9748-bfb5-47c1-8788-083b5189145c" containerID="d5fbfb18bc8dae146b5cbcdb5bea6e9020c81a674b5a478f89e75e8bf06f84ee" exitCode=0 Oct 06 09:59:31 crc kubenswrapper[4989]: I1006 09:59:31.597009 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"678e9748-bfb5-47c1-8788-083b5189145c","Type":"ContainerDied","Data":"d5fbfb18bc8dae146b5cbcdb5bea6e9020c81a674b5a478f89e75e8bf06f84ee"} Oct 06 09:59:31 crc kubenswrapper[4989]: I1006 09:59:31.601430 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"6b4c9853-f4c0-4e59-9d5a-78af6bf457b1","Type":"ContainerStarted","Data":"f4fd594385cf3e4f134c48be546da8305d81e6fa3cb3b2de56e5513a13d193be"} Oct 06 09:59:31 crc kubenswrapper[4989]: I1006 09:59:31.709089 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.709071492 podStartE2EDuration="7.709071492s" podCreationTimestamp="2025-10-06 09:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:31.699956549 +0000 UTC m=+4822.489982129" watchObservedRunningTime="2025-10-06 09:59:31.709071492 +0000 UTC m=+4822.499097062" Oct 06 09:59:32 crc kubenswrapper[4989]: I1006 09:59:32.615355 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"678e9748-bfb5-47c1-8788-083b5189145c","Type":"ContainerStarted","Data":"748944b5915ba15813cdbe9263ed2aef576f0d7c5f88589e5fb24d2a25bc9b8d"} Oct 06 09:59:32 crc kubenswrapper[4989]: I1006 09:59:32.658600 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.658565506 podStartE2EDuration="7.658565506s" podCreationTimestamp="2025-10-06 09:59:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 09:59:32.639824216 +0000 UTC m=+4823.429849827" watchObservedRunningTime="2025-10-06 09:59:32.658565506 +0000 UTC m=+4823.448591116" Oct 06 09:59:34 crc kubenswrapper[4989]: I1006 09:59:34.202798 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:34 crc kubenswrapper[4989]: I1006 09:59:34.359890 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 09:59:34 crc kubenswrapper[4989]: I1006 09:59:34.407293 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:34 crc kubenswrapper[4989]: I1006 09:59:34.633422 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="dnsmasq-dns" containerID="cri-o://90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f" gracePeriod=10 Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.077162 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.211717 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t82t7\" (UniqueName: \"kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7\") pod \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.211812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc\") pod \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.211879 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config\") pod \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\" (UID: \"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7\") " Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.219820 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7" (OuterVolumeSpecName: "kube-api-access-t82t7") pod "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" (UID: "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7"). InnerVolumeSpecName "kube-api-access-t82t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.274279 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config" (OuterVolumeSpecName: "config") pod "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" (UID: "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.274619 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" (UID: "bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.321853 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t82t7\" (UniqueName: \"kubernetes.io/projected/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-kube-api-access-t82t7\") on node \"crc\" DevicePath \"\"" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.322194 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.322320 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7-config\") on node \"crc\" DevicePath \"\"" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.642505 4989 generic.go:334] "Generic (PLEG): container finished" podID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerID="90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f" exitCode=0 Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.642565 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" event={"ID":"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7","Type":"ContainerDied","Data":"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f"} Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.642610 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.643041 4989 scope.go:117] "RemoveContainer" containerID="90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.643020 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-zkqcl" event={"ID":"bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7","Type":"ContainerDied","Data":"065a12502ced34bdb19e84f7c3048e4254ae14fa6f70bd31042668e14e97d49c"} Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.667239 4989 scope.go:117] "RemoveContainer" containerID="ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.681219 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.686121 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-zkqcl"] Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.697074 4989 scope.go:117] "RemoveContainer" containerID="90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f" Oct 06 09:59:35 crc kubenswrapper[4989]: E1006 09:59:35.697600 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f\": container with ID starting with 90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f not found: ID does not exist" containerID="90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.697638 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f"} err="failed to get container status \"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f\": rpc error: code = NotFound desc = could not find container \"90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f\": container with ID starting with 90ce5e7039c2986ffa4f873291341db02975ae8f7e5ad22c9ca737574b95f28f not found: ID does not exist" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.697737 4989 scope.go:117] "RemoveContainer" containerID="ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4" Oct 06 09:59:35 crc kubenswrapper[4989]: E1006 09:59:35.698245 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4\": container with ID starting with ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4 not found: ID does not exist" containerID="ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.698272 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4"} err="failed to get container status \"ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4\": rpc error: code = NotFound desc = could not find container \"ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4\": container with ID starting with ac23fb6663c2b89a9d25f44ccbc9fbfa11fd5f18943d6ffd8ae0064255f783c4 not found: ID does not exist" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.935928 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:59:35 crc kubenswrapper[4989]: E1006 09:59:35.936136 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.944887 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" path="/var/lib/kubelet/pods/bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7/volumes" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.976918 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:35 crc kubenswrapper[4989]: I1006 09:59:35.977002 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:36 crc kubenswrapper[4989]: I1006 09:59:36.377588 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 06 09:59:36 crc kubenswrapper[4989]: I1006 09:59:36.377728 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 06 09:59:38 crc kubenswrapper[4989]: I1006 09:59:38.048731 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:38 crc kubenswrapper[4989]: I1006 09:59:38.132587 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 06 09:59:38 crc kubenswrapper[4989]: I1006 09:59:38.456183 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 06 09:59:39 crc kubenswrapper[4989]: I1006 09:59:39.203143 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 06 09:59:47 crc kubenswrapper[4989]: I1006 09:59:47.937747 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:59:47 crc kubenswrapper[4989]: E1006 09:59:47.942413 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 09:59:59 crc kubenswrapper[4989]: I1006 09:59:59.944693 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 09:59:59 crc kubenswrapper[4989]: E1006 09:59:59.946877 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.169614 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94"] Oct 06 10:00:00 crc kubenswrapper[4989]: E1006 10:00:00.170576 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="init" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.170610 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="init" Oct 06 10:00:00 crc kubenswrapper[4989]: E1006 10:00:00.170634 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="dnsmasq-dns" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.170649 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="dnsmasq-dns" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.170985 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbb6dabe-38cf-4ed1-8012-97d5c17ec8c7" containerName="dnsmasq-dns" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.171847 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.175304 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.175905 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.184203 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94"] Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.263313 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.263386 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.263413 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxw2r\" (UniqueName: \"kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.364426 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.364504 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.364543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxw2r\" (UniqueName: \"kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.366526 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.378471 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.395612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxw2r\" (UniqueName: \"kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r\") pod \"collect-profiles-29329080-g4s94\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.562347 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.892875 4989 generic.go:334] "Generic (PLEG): container finished" podID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerID="13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b" exitCode=0 Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.892967 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerDied","Data":"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b"} Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.895527 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaaf8776-5729-4000-8e9d-153c983587cb" containerID="cf1cf42ad2c6da61ee8f7bb5a6ecda8e90f92656686681e24f2275818891c885" exitCode=0 Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.895581 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerDied","Data":"cf1cf42ad2c6da61ee8f7bb5a6ecda8e90f92656686681e24f2275818891c885"} Oct 06 10:00:00 crc kubenswrapper[4989]: I1006 10:00:00.993738 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94"] Oct 06 10:00:01 crc kubenswrapper[4989]: W1006 10:00:01.017849 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66463be1_c057_460f_96a2_1ed922b8621d.slice/crio-4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374 WatchSource:0}: Error finding container 4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374: Status 404 returned error can't find the container with id 4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374 Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.916125 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerStarted","Data":"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf"} Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.916846 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.921763 4989 generic.go:334] "Generic (PLEG): container finished" podID="66463be1-c057-460f-96a2-1ed922b8621d" containerID="676064bb353c3f0ca4e8ea4d0a568248559ef9d5fdb650ca5259948737657dd9" exitCode=0 Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.921897 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" event={"ID":"66463be1-c057-460f-96a2-1ed922b8621d","Type":"ContainerDied","Data":"676064bb353c3f0ca4e8ea4d0a568248559ef9d5fdb650ca5259948737657dd9"} Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.921954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" event={"ID":"66463be1-c057-460f-96a2-1ed922b8621d","Type":"ContainerStarted","Data":"4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374"} Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.924687 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerStarted","Data":"499eabd7dacca4f253dce95c20bc21ab46bb5a0fb615ed0bbe7f0836310c54fe"} Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.924914 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.952956 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.952930987 podStartE2EDuration="38.952930987s" podCreationTimestamp="2025-10-06 09:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:01.934781005 +0000 UTC m=+4852.724806615" watchObservedRunningTime="2025-10-06 10:00:01.952930987 +0000 UTC m=+4852.742956567" Oct 06 10:00:01 crc kubenswrapper[4989]: I1006 10:00:01.991881 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.991858519 podStartE2EDuration="37.991858519s" podCreationTimestamp="2025-10-06 09:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:01.984760734 +0000 UTC m=+4852.774786314" watchObservedRunningTime="2025-10-06 10:00:01.991858519 +0000 UTC m=+4852.781884099" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.241795 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.425182 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume\") pod \"66463be1-c057-460f-96a2-1ed922b8621d\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.425737 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxw2r\" (UniqueName: \"kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r\") pod \"66463be1-c057-460f-96a2-1ed922b8621d\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.425826 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume\") pod \"66463be1-c057-460f-96a2-1ed922b8621d\" (UID: \"66463be1-c057-460f-96a2-1ed922b8621d\") " Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.426219 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume" (OuterVolumeSpecName: "config-volume") pod "66463be1-c057-460f-96a2-1ed922b8621d" (UID: "66463be1-c057-460f-96a2-1ed922b8621d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.426452 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66463be1-c057-460f-96a2-1ed922b8621d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.440938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "66463be1-c057-460f-96a2-1ed922b8621d" (UID: "66463be1-c057-460f-96a2-1ed922b8621d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.441041 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r" (OuterVolumeSpecName: "kube-api-access-wxw2r") pod "66463be1-c057-460f-96a2-1ed922b8621d" (UID: "66463be1-c057-460f-96a2-1ed922b8621d"). InnerVolumeSpecName "kube-api-access-wxw2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.528165 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxw2r\" (UniqueName: \"kubernetes.io/projected/66463be1-c057-460f-96a2-1ed922b8621d-kube-api-access-wxw2r\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.528195 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66463be1-c057-460f-96a2-1ed922b8621d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.946853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" event={"ID":"66463be1-c057-460f-96a2-1ed922b8621d","Type":"ContainerDied","Data":"4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374"} Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.946894 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb6e6d2f890aaf71df101cdf0e055776068793f7c18b743d85e14dfd7ac1374" Oct 06 10:00:03 crc kubenswrapper[4989]: I1006 10:00:03.946948 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94" Oct 06 10:00:04 crc kubenswrapper[4989]: I1006 10:00:04.325555 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k"] Oct 06 10:00:04 crc kubenswrapper[4989]: I1006 10:00:04.330638 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329035-lsz9k"] Oct 06 10:00:05 crc kubenswrapper[4989]: I1006 10:00:05.954843 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72" path="/var/lib/kubelet/pods/b0e174d0-83b2-4fa2-8ca9-a0bc3c91cd72/volumes" Oct 06 10:00:10 crc kubenswrapper[4989]: I1006 10:00:10.936249 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 10:00:10 crc kubenswrapper[4989]: E1006 10:00:10.937733 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:00:15 crc kubenswrapper[4989]: I1006 10:00:15.295958 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 06 10:00:15 crc kubenswrapper[4989]: I1006 10:00:15.653886 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:19 crc kubenswrapper[4989]: I1006 10:00:19.808056 4989 scope.go:117] "RemoveContainer" containerID="b5382f952c8fad7bb2d79085bc023bbfaeb873c6af55a666a343e264c10d3f47" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.559495 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:00:20 crc kubenswrapper[4989]: E1006 10:00:20.560472 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66463be1-c057-460f-96a2-1ed922b8621d" containerName="collect-profiles" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.560500 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="66463be1-c057-460f-96a2-1ed922b8621d" containerName="collect-profiles" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.560762 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="66463be1-c057-460f-96a2-1ed922b8621d" containerName="collect-profiles" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.562503 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.578087 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.633482 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.633568 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwrvq\" (UniqueName: \"kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.633632 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.735283 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.735413 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.735465 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwrvq\" (UniqueName: \"kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.736765 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.737369 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.761891 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwrvq\" (UniqueName: \"kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq\") pod \"dnsmasq-dns-5b7946d7b9-f2bjf\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:20 crc kubenswrapper[4989]: I1006 10:00:20.950311 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:21 crc kubenswrapper[4989]: I1006 10:00:21.273605 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:00:21 crc kubenswrapper[4989]: W1006 10:00:21.282309 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9c8b79e_f897_49a3_8ce3_4d7630b55ae0.slice/crio-269f67ca234b82c77037ed594d9b258b2b6fa4e092e4248ede38378d86eb639e WatchSource:0}: Error finding container 269f67ca234b82c77037ed594d9b258b2b6fa4e092e4248ede38378d86eb639e: Status 404 returned error can't find the container with id 269f67ca234b82c77037ed594d9b258b2b6fa4e092e4248ede38378d86eb639e Oct 06 10:00:21 crc kubenswrapper[4989]: I1006 10:00:21.410629 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:22 crc kubenswrapper[4989]: I1006 10:00:22.142945 4989 generic.go:334] "Generic (PLEG): container finished" podID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerID="96862c2ec866b85a8607626cf1bc9890315efcd53d51ea77318825a1c03b2840" exitCode=0 Oct 06 10:00:22 crc kubenswrapper[4989]: I1006 10:00:22.143643 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" event={"ID":"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0","Type":"ContainerDied","Data":"96862c2ec866b85a8607626cf1bc9890315efcd53d51ea77318825a1c03b2840"} Oct 06 10:00:22 crc kubenswrapper[4989]: I1006 10:00:22.144260 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" event={"ID":"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0","Type":"ContainerStarted","Data":"269f67ca234b82c77037ed594d9b258b2b6fa4e092e4248ede38378d86eb639e"} Oct 06 10:00:22 crc kubenswrapper[4989]: I1006 10:00:22.350843 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:23 crc kubenswrapper[4989]: I1006 10:00:23.151678 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" event={"ID":"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0","Type":"ContainerStarted","Data":"5ad9d0aaa98c0398181a67bd203750809ea27633ee08b86391fd2084efbf8bed"} Oct 06 10:00:23 crc kubenswrapper[4989]: I1006 10:00:23.151894 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:23 crc kubenswrapper[4989]: I1006 10:00:23.174708 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" podStartSLOduration=3.174685764 podStartE2EDuration="3.174685764s" podCreationTimestamp="2025-10-06 10:00:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:00:23.169769892 +0000 UTC m=+4873.959795512" watchObservedRunningTime="2025-10-06 10:00:23.174685764 +0000 UTC m=+4873.964711354" Oct 06 10:00:23 crc kubenswrapper[4989]: I1006 10:00:23.227316 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="rabbitmq" containerID="cri-o://72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf" gracePeriod=604799 Oct 06 10:00:23 crc kubenswrapper[4989]: I1006 10:00:23.938051 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 10:00:23 crc kubenswrapper[4989]: E1006 10:00:23.938547 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:00:24 crc kubenswrapper[4989]: I1006 10:00:24.625128 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="rabbitmq" containerID="cri-o://499eabd7dacca4f253dce95c20bc21ab46bb5a0fb615ed0bbe7f0836310c54fe" gracePeriod=604798 Oct 06 10:00:25 crc kubenswrapper[4989]: I1006 10:00:25.294316 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.235:5672: connect: connection refused" Oct 06 10:00:25 crc kubenswrapper[4989]: I1006 10:00:25.651872 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.236:5672: connect: connection refused" Oct 06 10:00:29 crc kubenswrapper[4989]: I1006 10:00:29.878517 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.017976 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018296 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018359 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxhwj\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018430 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018497 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018576 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018701 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.018738 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info\") pod \"314237cb-dbf0-46d7-9830-6c1c10736eb8\" (UID: \"314237cb-dbf0-46d7-9830-6c1c10736eb8\") " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.020412 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.020450 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.020504 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.024789 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info" (OuterVolumeSpecName: "pod-info") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.028262 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj" (OuterVolumeSpecName: "kube-api-access-kxhwj") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "kube-api-access-kxhwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.041886 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.045622 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf" (OuterVolumeSpecName: "server-conf") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.047484 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead" (OuterVolumeSpecName: "persistence") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "pvc-2ef187cd-27c5-42fd-a438-81db725f4ead". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.114525 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "314237cb-dbf0-46d7-9830-6c1c10736eb8" (UID: "314237cb-dbf0-46d7-9830-6c1c10736eb8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120609 4989 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120717 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120819 4989 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/314237cb-dbf0-46d7-9830-6c1c10736eb8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120881 4989 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/314237cb-dbf0-46d7-9830-6c1c10736eb8-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120943 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.120998 4989 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/314237cb-dbf0-46d7-9830-6c1c10736eb8-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.121050 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.121145 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") on node \"crc\" " Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.121211 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxhwj\" (UniqueName: \"kubernetes.io/projected/314237cb-dbf0-46d7-9830-6c1c10736eb8-kube-api-access-kxhwj\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.144520 4989 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.144725 4989 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2ef187cd-27c5-42fd-a438-81db725f4ead" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead") on node "crc" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.221805 4989 generic.go:334] "Generic (PLEG): container finished" podID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerID="72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf" exitCode=0 Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.221864 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerDied","Data":"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf"} Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.221903 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"314237cb-dbf0-46d7-9830-6c1c10736eb8","Type":"ContainerDied","Data":"03d87d27a9096d18ed0c265b66eb3283360f5b51cdbd85ec98d99680a1a08d62"} Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.221930 4989 scope.go:117] "RemoveContainer" containerID="72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.222011 4989 reconciler_common.go:293] "Volume detached for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.222062 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.270815 4989 scope.go:117] "RemoveContainer" containerID="13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.313581 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.324612 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.336482 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:30 crc kubenswrapper[4989]: E1006 10:00:30.337064 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="setup-container" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.337094 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="setup-container" Oct 06 10:00:30 crc kubenswrapper[4989]: E1006 10:00:30.337103 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="rabbitmq" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.337109 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="rabbitmq" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.337318 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" containerName="rabbitmq" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.338324 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.342221 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.342712 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.342738 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x4jfh" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.342885 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.343117 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.343165 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.347516 4989 scope.go:117] "RemoveContainer" containerID="72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf" Oct 06 10:00:30 crc kubenswrapper[4989]: E1006 10:00:30.349450 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf\": container with ID starting with 72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf not found: ID does not exist" containerID="72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.349492 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf"} err="failed to get container status \"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf\": rpc error: code = NotFound desc = could not find container \"72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf\": container with ID starting with 72d0cf0436cf5df9c2a4f3c7a99b5c03a867b0f0274f19fbe0f62351312089cf not found: ID does not exist" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.349520 4989 scope.go:117] "RemoveContainer" containerID="13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b" Oct 06 10:00:30 crc kubenswrapper[4989]: E1006 10:00:30.350129 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b\": container with ID starting with 13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b not found: ID does not exist" containerID="13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.350162 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b"} err="failed to get container status \"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b\": rpc error: code = NotFound desc = could not find container \"13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b\": container with ID starting with 13f48be5ebd8f4e6cc644be6dbe3abeb7477a0611c21e1d918055fc1a18a146b not found: ID does not exist" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.526551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.526597 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.526626 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.526673 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df5fb509-bd0b-42b3-95d4-4a689e8713e5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.526896 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.527041 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df5fb509-bd0b-42b3-95d4-4a689e8713e5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.527077 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7jn5\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-kube-api-access-h7jn5\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.527144 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.527171 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.628848 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df5fb509-bd0b-42b3-95d4-4a689e8713e5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.628923 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.628976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df5fb509-bd0b-42b3-95d4-4a689e8713e5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629004 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7jn5\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-kube-api-access-h7jn5\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629037 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629058 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629090 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629118 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.629152 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.630075 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.630398 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.631189 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.631994 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.632055 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0927e9832013118ef8b604f038c0b9a415a2e415e922403e4d496bf8d6311ab9/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.632972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df5fb509-bd0b-42b3-95d4-4a689e8713e5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.634403 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.637480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df5fb509-bd0b-42b3-95d4-4a689e8713e5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.643972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df5fb509-bd0b-42b3-95d4-4a689e8713e5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.660225 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7jn5\" (UniqueName: \"kubernetes.io/projected/df5fb509-bd0b-42b3-95d4-4a689e8713e5-kube-api-access-h7jn5\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.679813 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2ef187cd-27c5-42fd-a438-81db725f4ead\") pod \"rabbitmq-server-0\" (UID: \"df5fb509-bd0b-42b3-95d4-4a689e8713e5\") " pod="openstack/rabbitmq-server-0" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.952061 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:00:30 crc kubenswrapper[4989]: I1006 10:00:30.961971 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.042229 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.042436 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="dnsmasq-dns" containerID="cri-o://19844a6861d6f0a7afd3a0081319ef88bed9fc7e2b76b300c54e7422ea204681" gracePeriod=10 Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.247083 4989 generic.go:334] "Generic (PLEG): container finished" podID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerID="19844a6861d6f0a7afd3a0081319ef88bed9fc7e2b76b300c54e7422ea204681" exitCode=0 Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.247544 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" event={"ID":"480bbde2-3ca3-4e6f-bc34-e530774cb253","Type":"ContainerDied","Data":"19844a6861d6f0a7afd3a0081319ef88bed9fc7e2b76b300c54e7422ea204681"} Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.250389 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.252712 4989 generic.go:334] "Generic (PLEG): container finished" podID="aaaf8776-5729-4000-8e9d-153c983587cb" containerID="499eabd7dacca4f253dce95c20bc21ab46bb5a0fb615ed0bbe7f0836310c54fe" exitCode=0 Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.252755 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerDied","Data":"499eabd7dacca4f253dce95c20bc21ab46bb5a0fb615ed0bbe7f0836310c54fe"} Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.252810 4989 scope.go:117] "RemoveContainer" containerID="499eabd7dacca4f253dce95c20bc21ab46bb5a0fb615ed0bbe7f0836310c54fe" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.289841 4989 scope.go:117] "RemoveContainer" containerID="cf1cf42ad2c6da61ee8f7bb5a6ecda8e90f92656686681e24f2275818891c885" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344538 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344601 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344647 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344718 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344745 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344773 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344815 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwsmf\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.344868 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret\") pod \"aaaf8776-5729-4000-8e9d-153c983587cb\" (UID: \"aaaf8776-5729-4000-8e9d-153c983587cb\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.345836 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.346299 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.347151 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.351549 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.351735 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf" (OuterVolumeSpecName: "kube-api-access-zwsmf") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "kube-api-access-zwsmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.359783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e" (OuterVolumeSpecName: "persistence") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "pvc-584be944-0069-4476-93e5-0c28a610bf0e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.362071 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info" (OuterVolumeSpecName: "pod-info") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.369810 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf" (OuterVolumeSpecName: "server-conf") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446016 4989 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-server-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446265 4989 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aaaf8776-5729-4000-8e9d-153c983587cb-pod-info\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446281 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwsmf\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-kube-api-access-zwsmf\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446291 4989 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aaaf8776-5729-4000-8e9d-153c983587cb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446313 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") on node \"crc\" " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446205 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "aaaf8776-5729-4000-8e9d-153c983587cb" (UID: "aaaf8776-5729-4000-8e9d-153c983587cb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446323 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446391 4989 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aaaf8776-5729-4000-8e9d-153c983587cb-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.446410 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.464544 4989 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.464849 4989 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-584be944-0069-4476-93e5-0c28a610bf0e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e") on node "crc" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.534637 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.548852 4989 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aaaf8776-5729-4000-8e9d-153c983587cb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.548890 4989 reconciler_common.go:293] "Volume detached for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.556792 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.649740 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc\") pod \"480bbde2-3ca3-4e6f-bc34-e530774cb253\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.649892 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bktln\" (UniqueName: \"kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln\") pod \"480bbde2-3ca3-4e6f-bc34-e530774cb253\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.649960 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config\") pod \"480bbde2-3ca3-4e6f-bc34-e530774cb253\" (UID: \"480bbde2-3ca3-4e6f-bc34-e530774cb253\") " Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.654037 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln" (OuterVolumeSpecName: "kube-api-access-bktln") pod "480bbde2-3ca3-4e6f-bc34-e530774cb253" (UID: "480bbde2-3ca3-4e6f-bc34-e530774cb253"). InnerVolumeSpecName "kube-api-access-bktln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.689769 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "480bbde2-3ca3-4e6f-bc34-e530774cb253" (UID: "480bbde2-3ca3-4e6f-bc34-e530774cb253"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.693898 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config" (OuterVolumeSpecName: "config") pod "480bbde2-3ca3-4e6f-bc34-e530774cb253" (UID: "480bbde2-3ca3-4e6f-bc34-e530774cb253"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.751941 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.751983 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bktln\" (UniqueName: \"kubernetes.io/projected/480bbde2-3ca3-4e6f-bc34-e530774cb253-kube-api-access-bktln\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.751997 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/480bbde2-3ca3-4e6f-bc34-e530774cb253-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:00:31 crc kubenswrapper[4989]: I1006 10:00:31.945216 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="314237cb-dbf0-46d7-9830-6c1c10736eb8" path="/var/lib/kubelet/pods/314237cb-dbf0-46d7-9830-6c1c10736eb8/volumes" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.263524 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.263779 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-d5jbr" event={"ID":"480bbde2-3ca3-4e6f-bc34-e530774cb253","Type":"ContainerDied","Data":"af181ac88906bda571613a6a3f23400ff95d84d0516de748928d6c0fece47bb1"} Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.264479 4989 scope.go:117] "RemoveContainer" containerID="19844a6861d6f0a7afd3a0081319ef88bed9fc7e2b76b300c54e7422ea204681" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.266978 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"aaaf8776-5729-4000-8e9d-153c983587cb","Type":"ContainerDied","Data":"15c00c981afabd675a6e33f0d65e1f75ebc04db345659242659f5c27012fa07c"} Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.267613 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.269727 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df5fb509-bd0b-42b3-95d4-4a689e8713e5","Type":"ContainerStarted","Data":"9c29a00fd53c3bce2926ace3e5eae1924b2904e266c7f9814d7a32bf87023f01"} Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.301198 4989 scope.go:117] "RemoveContainer" containerID="4f9b4b3a4072da73ea2b0a384dbf5eec6fb4d729c64c721b2d7d2af75772f274" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.304398 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.330617 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.341190 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.369883 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-d5jbr"] Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.380473 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:32 crc kubenswrapper[4989]: E1006 10:00:32.380979 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="setup-container" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381002 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="setup-container" Oct 06 10:00:32 crc kubenswrapper[4989]: E1006 10:00:32.381046 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="dnsmasq-dns" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381057 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="dnsmasq-dns" Oct 06 10:00:32 crc kubenswrapper[4989]: E1006 10:00:32.381085 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="init" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381092 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="init" Oct 06 10:00:32 crc kubenswrapper[4989]: E1006 10:00:32.381132 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="rabbitmq" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381143 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="rabbitmq" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381459 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" containerName="dnsmasq-dns" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.381497 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" containerName="rabbitmq" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.382905 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.385758 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hx85m" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.385783 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.385899 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.388327 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.389136 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.404001 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.564639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.564757 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.564835 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.564881 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.565011 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.565085 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.565133 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.565168 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g5c6\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-kube-api-access-5g5c6\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.565249 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.667512 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.667629 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.667697 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.667795 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.668606 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.668814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.668913 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.668952 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g5c6\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-kube-api-access-5g5c6\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.669021 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.669233 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.669567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.669894 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.670151 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.675915 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.675968 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/21bc325a0270dfafae871c98be722654f90969b8795d181372052a53810d41e3/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.676155 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.680543 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.682365 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.689074 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g5c6\" (UniqueName: \"kubernetes.io/projected/4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf-kube-api-access-5g5c6\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:32 crc kubenswrapper[4989]: I1006 10:00:32.712332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-584be944-0069-4476-93e5-0c28a610bf0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-584be944-0069-4476-93e5-0c28a610bf0e\") pod \"rabbitmq-cell1-server-0\" (UID: \"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf\") " pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:33 crc kubenswrapper[4989]: I1006 10:00:33.002031 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:00:33 crc kubenswrapper[4989]: I1006 10:00:33.282743 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df5fb509-bd0b-42b3-95d4-4a689e8713e5","Type":"ContainerStarted","Data":"e1b0fee7da3e8d4a75247a266c8d37a9d06a8aac6226ab8dd9e18f768498fc58"} Oct 06 10:00:33 crc kubenswrapper[4989]: I1006 10:00:33.515456 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 06 10:00:33 crc kubenswrapper[4989]: I1006 10:00:33.960153 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480bbde2-3ca3-4e6f-bc34-e530774cb253" path="/var/lib/kubelet/pods/480bbde2-3ca3-4e6f-bc34-e530774cb253/volumes" Oct 06 10:00:33 crc kubenswrapper[4989]: I1006 10:00:33.961684 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaaf8776-5729-4000-8e9d-153c983587cb" path="/var/lib/kubelet/pods/aaaf8776-5729-4000-8e9d-153c983587cb/volumes" Oct 06 10:00:34 crc kubenswrapper[4989]: I1006 10:00:34.296379 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf","Type":"ContainerStarted","Data":"ae957ff0f63101e4c4072443559158ef434c8f8f8708240fdc4cdf4ec70bba18"} Oct 06 10:00:35 crc kubenswrapper[4989]: I1006 10:00:35.306173 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf","Type":"ContainerStarted","Data":"5dbbceb54ddd27125f8e3482060568f4d2a95fde00e68e27f0ee96651bf33baa"} Oct 06 10:00:38 crc kubenswrapper[4989]: I1006 10:00:38.936565 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 10:00:39 crc kubenswrapper[4989]: I1006 10:00:39.351136 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b"} Oct 06 10:01:06 crc kubenswrapper[4989]: I1006 10:01:06.662238 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df5fb509-bd0b-42b3-95d4-4a689e8713e5","Type":"ContainerDied","Data":"e1b0fee7da3e8d4a75247a266c8d37a9d06a8aac6226ab8dd9e18f768498fc58"} Oct 06 10:01:06 crc kubenswrapper[4989]: I1006 10:01:06.663842 4989 generic.go:334] "Generic (PLEG): container finished" podID="df5fb509-bd0b-42b3-95d4-4a689e8713e5" containerID="e1b0fee7da3e8d4a75247a266c8d37a9d06a8aac6226ab8dd9e18f768498fc58" exitCode=0 Oct 06 10:01:07 crc kubenswrapper[4989]: I1006 10:01:07.675208 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"df5fb509-bd0b-42b3-95d4-4a689e8713e5","Type":"ContainerStarted","Data":"2aa5be5b70611e60f4bd4fe0f44e592fc94b4656168442e3a302b613f6d1863e"} Oct 06 10:01:07 crc kubenswrapper[4989]: I1006 10:01:07.675993 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 06 10:01:07 crc kubenswrapper[4989]: I1006 10:01:07.722577 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.722544269 podStartE2EDuration="37.722544269s" podCreationTimestamp="2025-10-06 10:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:01:07.705811327 +0000 UTC m=+4918.495836967" watchObservedRunningTime="2025-10-06 10:01:07.722544269 +0000 UTC m=+4918.512569889" Oct 06 10:01:08 crc kubenswrapper[4989]: I1006 10:01:08.688975 4989 generic.go:334] "Generic (PLEG): container finished" podID="4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf" containerID="5dbbceb54ddd27125f8e3482060568f4d2a95fde00e68e27f0ee96651bf33baa" exitCode=0 Oct 06 10:01:08 crc kubenswrapper[4989]: I1006 10:01:08.689062 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf","Type":"ContainerDied","Data":"5dbbceb54ddd27125f8e3482060568f4d2a95fde00e68e27f0ee96651bf33baa"} Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.191178 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.193921 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.205017 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.358982 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.359082 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.359351 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nc4p\" (UniqueName: \"kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.461930 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.462006 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.462047 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nc4p\" (UniqueName: \"kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.462910 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.463001 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.583291 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nc4p\" (UniqueName: \"kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p\") pod \"redhat-operators-hgmdd\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.703160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf","Type":"ContainerStarted","Data":"a36a47299a4fd41828ad856a5c1612c5118e2672e7463603d31249716c17bde7"} Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.703477 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.743285 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.743258227 podStartE2EDuration="37.743258227s" podCreationTimestamp="2025-10-06 10:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:01:09.733905238 +0000 UTC m=+4920.523930808" watchObservedRunningTime="2025-10-06 10:01:09.743258227 +0000 UTC m=+4920.533283817" Oct 06 10:01:09 crc kubenswrapper[4989]: I1006 10:01:09.831411 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:10 crc kubenswrapper[4989]: I1006 10:01:10.303902 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:10 crc kubenswrapper[4989]: I1006 10:01:10.730279 4989 generic.go:334] "Generic (PLEG): container finished" podID="a57b40da-709a-4d4a-b950-61834badc6ea" containerID="05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570" exitCode=0 Oct 06 10:01:10 crc kubenswrapper[4989]: I1006 10:01:10.730369 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerDied","Data":"05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570"} Oct 06 10:01:10 crc kubenswrapper[4989]: I1006 10:01:10.730769 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerStarted","Data":"e995dbd6c4448d7da41ad006648daf3742abd4e515a1b3500518cee14e75d8c6"} Oct 06 10:01:10 crc kubenswrapper[4989]: I1006 10:01:10.732321 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:01:12 crc kubenswrapper[4989]: I1006 10:01:12.757061 4989 generic.go:334] "Generic (PLEG): container finished" podID="a57b40da-709a-4d4a-b950-61834badc6ea" containerID="6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af" exitCode=0 Oct 06 10:01:12 crc kubenswrapper[4989]: I1006 10:01:12.757192 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerDied","Data":"6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af"} Oct 06 10:01:13 crc kubenswrapper[4989]: I1006 10:01:13.772617 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerStarted","Data":"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2"} Oct 06 10:01:13 crc kubenswrapper[4989]: I1006 10:01:13.811596 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hgmdd" podStartSLOduration=2.304543174 podStartE2EDuration="4.811571689s" podCreationTimestamp="2025-10-06 10:01:09 +0000 UTC" firstStartedPulling="2025-10-06 10:01:10.732114258 +0000 UTC m=+4921.522139838" lastFinishedPulling="2025-10-06 10:01:13.239142733 +0000 UTC m=+4924.029168353" observedRunningTime="2025-10-06 10:01:13.80258155 +0000 UTC m=+4924.592607130" watchObservedRunningTime="2025-10-06 10:01:13.811571689 +0000 UTC m=+4924.601597269" Oct 06 10:01:19 crc kubenswrapper[4989]: I1006 10:01:19.832582 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:19 crc kubenswrapper[4989]: I1006 10:01:19.833542 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:19 crc kubenswrapper[4989]: I1006 10:01:19.923621 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:20 crc kubenswrapper[4989]: I1006 10:01:20.022978 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:20 crc kubenswrapper[4989]: I1006 10:01:20.178009 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:20 crc kubenswrapper[4989]: I1006 10:01:20.968811 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 06 10:01:21 crc kubenswrapper[4989]: I1006 10:01:21.875957 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hgmdd" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="registry-server" containerID="cri-o://857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2" gracePeriod=2 Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.755981 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.893401 4989 generic.go:334] "Generic (PLEG): container finished" podID="a57b40da-709a-4d4a-b950-61834badc6ea" containerID="857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2" exitCode=0 Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.893493 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerDied","Data":"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2"} Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.893520 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hgmdd" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.893556 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hgmdd" event={"ID":"a57b40da-709a-4d4a-b950-61834badc6ea","Type":"ContainerDied","Data":"e995dbd6c4448d7da41ad006648daf3742abd4e515a1b3500518cee14e75d8c6"} Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.893598 4989 scope.go:117] "RemoveContainer" containerID="857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.894638 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities\") pod \"a57b40da-709a-4d4a-b950-61834badc6ea\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.894985 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nc4p\" (UniqueName: \"kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p\") pod \"a57b40da-709a-4d4a-b950-61834badc6ea\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.895103 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content\") pod \"a57b40da-709a-4d4a-b950-61834badc6ea\" (UID: \"a57b40da-709a-4d4a-b950-61834badc6ea\") " Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.896597 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities" (OuterVolumeSpecName: "utilities") pod "a57b40da-709a-4d4a-b950-61834badc6ea" (UID: "a57b40da-709a-4d4a-b950-61834badc6ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.907458 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p" (OuterVolumeSpecName: "kube-api-access-6nc4p") pod "a57b40da-709a-4d4a-b950-61834badc6ea" (UID: "a57b40da-709a-4d4a-b950-61834badc6ea"). InnerVolumeSpecName "kube-api-access-6nc4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.959478 4989 scope.go:117] "RemoveContainer" containerID="6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.998229 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:22 crc kubenswrapper[4989]: I1006 10:01:22.998279 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nc4p\" (UniqueName: \"kubernetes.io/projected/a57b40da-709a-4d4a-b950-61834badc6ea-kube-api-access-6nc4p\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.003025 4989 scope.go:117] "RemoveContainer" containerID="05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.009021 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.028351 4989 scope.go:117] "RemoveContainer" containerID="857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2" Oct 06 10:01:23 crc kubenswrapper[4989]: E1006 10:01:23.032523 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2\": container with ID starting with 857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2 not found: ID does not exist" containerID="857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.032589 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2"} err="failed to get container status \"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2\": rpc error: code = NotFound desc = could not find container \"857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2\": container with ID starting with 857066449d4c9d1244e8c64dc7ce7b133cceab74574000db8b5901b283ac8fb2 not found: ID does not exist" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.032631 4989 scope.go:117] "RemoveContainer" containerID="6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af" Oct 06 10:01:23 crc kubenswrapper[4989]: E1006 10:01:23.036995 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af\": container with ID starting with 6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af not found: ID does not exist" containerID="6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.037062 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af"} err="failed to get container status \"6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af\": rpc error: code = NotFound desc = could not find container \"6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af\": container with ID starting with 6ce62d689ec4d8e847f879a273f2a809770d0e78774cd483c305301ea10a31af not found: ID does not exist" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.037105 4989 scope.go:117] "RemoveContainer" containerID="05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570" Oct 06 10:01:23 crc kubenswrapper[4989]: E1006 10:01:23.038085 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570\": container with ID starting with 05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570 not found: ID does not exist" containerID="05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.038123 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570"} err="failed to get container status \"05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570\": rpc error: code = NotFound desc = could not find container \"05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570\": container with ID starting with 05e83e03fbdbac26c1e5f93793c2df861c59e9f33e65520c060fd5bc6f62e570 not found: ID does not exist" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.417281 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a57b40da-709a-4d4a-b950-61834badc6ea" (UID: "a57b40da-709a-4d4a-b950-61834badc6ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.508971 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a57b40da-709a-4d4a-b950-61834badc6ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.555706 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.567280 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hgmdd"] Oct 06 10:01:23 crc kubenswrapper[4989]: I1006 10:01:23.949350 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" path="/var/lib/kubelet/pods/a57b40da-709a-4d4a-b950-61834badc6ea/volumes" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.019727 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:29 crc kubenswrapper[4989]: E1006 10:01:29.022958 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="registry-server" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.022996 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="registry-server" Oct 06 10:01:29 crc kubenswrapper[4989]: E1006 10:01:29.023027 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="extract-utilities" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.023040 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="extract-utilities" Oct 06 10:01:29 crc kubenswrapper[4989]: E1006 10:01:29.023098 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="extract-content" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.023111 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="extract-content" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.023415 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57b40da-709a-4d4a-b950-61834badc6ea" containerName="registry-server" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.025422 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.032899 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.135030 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.135406 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.135470 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4mp\" (UniqueName: \"kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.237300 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.237416 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.237501 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4mp\" (UniqueName: \"kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.238340 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.238575 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.275731 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4mp\" (UniqueName: \"kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp\") pod \"certified-operators-dk2nj\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:29 crc kubenswrapper[4989]: I1006 10:01:29.363328 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:30 crc kubenswrapper[4989]: I1006 10:01:29.950670 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:30 crc kubenswrapper[4989]: I1006 10:01:29.981347 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerStarted","Data":"003fd98c209d40b614bebd1a4ed7f5f1401d682b9848db7a9f1798c17b676ea5"} Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.002166 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerID="77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575" exitCode=0 Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.002608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerDied","Data":"77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575"} Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.255565 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.260758 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.265175 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.272303 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.399060 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqdsb\" (UniqueName: \"kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb\") pod \"mariadb-client-1-default\" (UID: \"f47b11b4-1757-466d-abd9-fda90c0ab198\") " pod="openstack/mariadb-client-1-default" Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.501145 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqdsb\" (UniqueName: \"kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb\") pod \"mariadb-client-1-default\" (UID: \"f47b11b4-1757-466d-abd9-fda90c0ab198\") " pod="openstack/mariadb-client-1-default" Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.539231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqdsb\" (UniqueName: \"kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb\") pod \"mariadb-client-1-default\" (UID: \"f47b11b4-1757-466d-abd9-fda90c0ab198\") " pod="openstack/mariadb-client-1-default" Oct 06 10:01:31 crc kubenswrapper[4989]: I1006 10:01:31.599922 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 06 10:01:32 crc kubenswrapper[4989]: I1006 10:01:32.019486 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerStarted","Data":"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12"} Oct 06 10:01:32 crc kubenswrapper[4989]: I1006 10:01:32.294085 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 06 10:01:32 crc kubenswrapper[4989]: W1006 10:01:32.300905 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf47b11b4_1757_466d_abd9_fda90c0ab198.slice/crio-16f952ff32b52594d3be0108eb6d103232a1560759e0bff6449585212a8aa940 WatchSource:0}: Error finding container 16f952ff32b52594d3be0108eb6d103232a1560759e0bff6449585212a8aa940: Status 404 returned error can't find the container with id 16f952ff32b52594d3be0108eb6d103232a1560759e0bff6449585212a8aa940 Oct 06 10:01:33 crc kubenswrapper[4989]: I1006 10:01:33.035208 4989 generic.go:334] "Generic (PLEG): container finished" podID="f47b11b4-1757-466d-abd9-fda90c0ab198" containerID="d9eaa9f17cb69081351d566ddc882fff1a56a760bb9adde010a3aa60706b6d73" exitCode=0 Oct 06 10:01:33 crc kubenswrapper[4989]: I1006 10:01:33.035361 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"f47b11b4-1757-466d-abd9-fda90c0ab198","Type":"ContainerDied","Data":"d9eaa9f17cb69081351d566ddc882fff1a56a760bb9adde010a3aa60706b6d73"} Oct 06 10:01:33 crc kubenswrapper[4989]: I1006 10:01:33.035754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"f47b11b4-1757-466d-abd9-fda90c0ab198","Type":"ContainerStarted","Data":"16f952ff32b52594d3be0108eb6d103232a1560759e0bff6449585212a8aa940"} Oct 06 10:01:33 crc kubenswrapper[4989]: I1006 10:01:33.038643 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerID="ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12" exitCode=0 Oct 06 10:01:33 crc kubenswrapper[4989]: I1006 10:01:33.038698 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerDied","Data":"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12"} Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.048890 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerStarted","Data":"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143"} Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.078761 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dk2nj" podStartSLOduration=3.519233816 podStartE2EDuration="6.078731823s" podCreationTimestamp="2025-10-06 10:01:28 +0000 UTC" firstStartedPulling="2025-10-06 10:01:31.008468086 +0000 UTC m=+4941.798493716" lastFinishedPulling="2025-10-06 10:01:33.567966103 +0000 UTC m=+4944.357991723" observedRunningTime="2025-10-06 10:01:34.075577632 +0000 UTC m=+4944.865603212" watchObservedRunningTime="2025-10-06 10:01:34.078731823 +0000 UTC m=+4944.868757403" Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.492860 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.526432 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_f47b11b4-1757-466d-abd9-fda90c0ab198/mariadb-client-1-default/0.log" Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.556574 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.564557 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.564628 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqdsb\" (UniqueName: \"kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb\") pod \"f47b11b4-1757-466d-abd9-fda90c0ab198\" (UID: \"f47b11b4-1757-466d-abd9-fda90c0ab198\") " Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.571587 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb" (OuterVolumeSpecName: "kube-api-access-pqdsb") pod "f47b11b4-1757-466d-abd9-fda90c0ab198" (UID: "f47b11b4-1757-466d-abd9-fda90c0ab198"). InnerVolumeSpecName "kube-api-access-pqdsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:34 crc kubenswrapper[4989]: I1006 10:01:34.667096 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqdsb\" (UniqueName: \"kubernetes.io/projected/f47b11b4-1757-466d-abd9-fda90c0ab198-kube-api-access-pqdsb\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.067914 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.067919 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16f952ff32b52594d3be0108eb6d103232a1560759e0bff6449585212a8aa940" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.106676 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 06 10:01:35 crc kubenswrapper[4989]: E1006 10:01:35.108287 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47b11b4-1757-466d-abd9-fda90c0ab198" containerName="mariadb-client-1-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.108320 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47b11b4-1757-466d-abd9-fda90c0ab198" containerName="mariadb-client-1-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.108625 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f47b11b4-1757-466d-abd9-fda90c0ab198" containerName="mariadb-client-1-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.109514 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.114554 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.146547 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.176376 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdjs\" (UniqueName: \"kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs\") pod \"mariadb-client-2-default\" (UID: \"6181b829-fa68-4141-b3c6-b84b85aef343\") " pod="openstack/mariadb-client-2-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.278767 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdjs\" (UniqueName: \"kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs\") pod \"mariadb-client-2-default\" (UID: \"6181b829-fa68-4141-b3c6-b84b85aef343\") " pod="openstack/mariadb-client-2-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.308276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdjs\" (UniqueName: \"kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs\") pod \"mariadb-client-2-default\" (UID: \"6181b829-fa68-4141-b3c6-b84b85aef343\") " pod="openstack/mariadb-client-2-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.431780 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.872958 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 06 10:01:35 crc kubenswrapper[4989]: W1006 10:01:35.883228 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6181b829_fa68_4141_b3c6_b84b85aef343.slice/crio-1ee45d1b8639968345f31ef65a4a81d7ab1e9ac91e62c0d0b2c745ec5524cc5e WatchSource:0}: Error finding container 1ee45d1b8639968345f31ef65a4a81d7ab1e9ac91e62c0d0b2c745ec5524cc5e: Status 404 returned error can't find the container with id 1ee45d1b8639968345f31ef65a4a81d7ab1e9ac91e62c0d0b2c745ec5524cc5e Oct 06 10:01:35 crc kubenswrapper[4989]: I1006 10:01:35.951392 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f47b11b4-1757-466d-abd9-fda90c0ab198" path="/var/lib/kubelet/pods/f47b11b4-1757-466d-abd9-fda90c0ab198/volumes" Oct 06 10:01:36 crc kubenswrapper[4989]: I1006 10:01:36.094418 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"6181b829-fa68-4141-b3c6-b84b85aef343","Type":"ContainerStarted","Data":"1ee45d1b8639968345f31ef65a4a81d7ab1e9ac91e62c0d0b2c745ec5524cc5e"} Oct 06 10:01:36 crc kubenswrapper[4989]: I1006 10:01:36.116139 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.116111062 podStartE2EDuration="1.116111062s" podCreationTimestamp="2025-10-06 10:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:01:36.111474339 +0000 UTC m=+4946.901499939" watchObservedRunningTime="2025-10-06 10:01:36.116111062 +0000 UTC m=+4946.906136652" Oct 06 10:01:37 crc kubenswrapper[4989]: I1006 10:01:37.109316 4989 generic.go:334] "Generic (PLEG): container finished" podID="6181b829-fa68-4141-b3c6-b84b85aef343" containerID="28dabe457ba3c772d3e1b22e40dc182bb317356ae9a7fb7a66dbfe7373ba5301" exitCode=0 Oct 06 10:01:37 crc kubenswrapper[4989]: I1006 10:01:37.109711 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"6181b829-fa68-4141-b3c6-b84b85aef343","Type":"ContainerDied","Data":"28dabe457ba3c772d3e1b22e40dc182bb317356ae9a7fb7a66dbfe7373ba5301"} Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.669295 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.714406 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.725216 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.747095 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqdjs\" (UniqueName: \"kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs\") pod \"6181b829-fa68-4141-b3c6-b84b85aef343\" (UID: \"6181b829-fa68-4141-b3c6-b84b85aef343\") " Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.762219 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs" (OuterVolumeSpecName: "kube-api-access-bqdjs") pod "6181b829-fa68-4141-b3c6-b84b85aef343" (UID: "6181b829-fa68-4141-b3c6-b84b85aef343"). InnerVolumeSpecName "kube-api-access-bqdjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:38 crc kubenswrapper[4989]: I1006 10:01:38.848959 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqdjs\" (UniqueName: \"kubernetes.io/projected/6181b829-fa68-4141-b3c6-b84b85aef343-kube-api-access-bqdjs\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.131355 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee45d1b8639968345f31ef65a4a81d7ab1e9ac91e62c0d0b2c745ec5524cc5e" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.131543 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.222526 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 06 10:01:39 crc kubenswrapper[4989]: E1006 10:01:39.222899 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6181b829-fa68-4141-b3c6-b84b85aef343" containerName="mariadb-client-2-default" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.222916 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6181b829-fa68-4141-b3c6-b84b85aef343" containerName="mariadb-client-2-default" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.223130 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6181b829-fa68-4141-b3c6-b84b85aef343" containerName="mariadb-client-2-default" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.223805 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.229415 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.244430 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.256173 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmnl4\" (UniqueName: \"kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4\") pod \"mariadb-client-1\" (UID: \"c715de11-638e-41dc-8df9-adfb4f28acbb\") " pod="openstack/mariadb-client-1" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.358850 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmnl4\" (UniqueName: \"kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4\") pod \"mariadb-client-1\" (UID: \"c715de11-638e-41dc-8df9-adfb4f28acbb\") " pod="openstack/mariadb-client-1" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.364494 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.364551 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.383331 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmnl4\" (UniqueName: \"kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4\") pod \"mariadb-client-1\" (UID: \"c715de11-638e-41dc-8df9-adfb4f28acbb\") " pod="openstack/mariadb-client-1" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.455279 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.562759 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 06 10:01:39 crc kubenswrapper[4989]: I1006 10:01:39.953235 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6181b829-fa68-4141-b3c6-b84b85aef343" path="/var/lib/kubelet/pods/6181b829-fa68-4141-b3c6-b84b85aef343/volumes" Oct 06 10:01:40 crc kubenswrapper[4989]: I1006 10:01:40.199496 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 06 10:01:40 crc kubenswrapper[4989]: W1006 10:01:40.216260 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc715de11_638e_41dc_8df9_adfb4f28acbb.slice/crio-806a400ee3481962ca8d84ae2c8fcf07bc8cb4f684fe9dff9bb341ef1146b25a WatchSource:0}: Error finding container 806a400ee3481962ca8d84ae2c8fcf07bc8cb4f684fe9dff9bb341ef1146b25a: Status 404 returned error can't find the container with id 806a400ee3481962ca8d84ae2c8fcf07bc8cb4f684fe9dff9bb341ef1146b25a Oct 06 10:01:40 crc kubenswrapper[4989]: I1006 10:01:40.238662 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:40 crc kubenswrapper[4989]: I1006 10:01:40.397755 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:41 crc kubenswrapper[4989]: I1006 10:01:41.158092 4989 generic.go:334] "Generic (PLEG): container finished" podID="c715de11-638e-41dc-8df9-adfb4f28acbb" containerID="035e02d2fe942373e5d3d40748d687003a90c7fec19f1a6def891a41dfbb8386" exitCode=0 Oct 06 10:01:41 crc kubenswrapper[4989]: I1006 10:01:41.158209 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"c715de11-638e-41dc-8df9-adfb4f28acbb","Type":"ContainerDied","Data":"035e02d2fe942373e5d3d40748d687003a90c7fec19f1a6def891a41dfbb8386"} Oct 06 10:01:41 crc kubenswrapper[4989]: I1006 10:01:41.158300 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"c715de11-638e-41dc-8df9-adfb4f28acbb","Type":"ContainerStarted","Data":"806a400ee3481962ca8d84ae2c8fcf07bc8cb4f684fe9dff9bb341ef1146b25a"} Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.168907 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dk2nj" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="registry-server" containerID="cri-o://86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143" gracePeriod=2 Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.808193 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.817110 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.829390 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities\") pod \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.830804 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content\") pod \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.830870 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp4mp\" (UniqueName: \"kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp\") pod \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\" (UID: \"0a1cad76-c7a2-4331-bcbe-8964c1c06b41\") " Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.830914 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmnl4\" (UniqueName: \"kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4\") pod \"c715de11-638e-41dc-8df9-adfb4f28acbb\" (UID: \"c715de11-638e-41dc-8df9-adfb4f28acbb\") " Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.845393 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities" (OuterVolumeSpecName: "utilities") pod "0a1cad76-c7a2-4331-bcbe-8964c1c06b41" (UID: "0a1cad76-c7a2-4331-bcbe-8964c1c06b41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.867990 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4" (OuterVolumeSpecName: "kube-api-access-rmnl4") pod "c715de11-638e-41dc-8df9-adfb4f28acbb" (UID: "c715de11-638e-41dc-8df9-adfb4f28acbb"). InnerVolumeSpecName "kube-api-access-rmnl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.887391 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_c715de11-638e-41dc-8df9-adfb4f28acbb/mariadb-client-1/0.log" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.887938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp" (OuterVolumeSpecName: "kube-api-access-wp4mp") pod "0a1cad76-c7a2-4331-bcbe-8964c1c06b41" (UID: "0a1cad76-c7a2-4331-bcbe-8964c1c06b41"). InnerVolumeSpecName "kube-api-access-wp4mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.926048 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.931489 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.932870 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a1cad76-c7a2-4331-bcbe-8964c1c06b41" (UID: "0a1cad76-c7a2-4331-bcbe-8964c1c06b41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.933326 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp4mp\" (UniqueName: \"kubernetes.io/projected/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-kube-api-access-wp4mp\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.933362 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmnl4\" (UniqueName: \"kubernetes.io/projected/c715de11-638e-41dc-8df9-adfb4f28acbb-kube-api-access-rmnl4\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.933375 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:42 crc kubenswrapper[4989]: I1006 10:01:42.933387 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a1cad76-c7a2-4331-bcbe-8964c1c06b41-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.189000 4989 generic.go:334] "Generic (PLEG): container finished" podID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerID="86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143" exitCode=0 Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.189135 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerDied","Data":"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143"} Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.189184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dk2nj" event={"ID":"0a1cad76-c7a2-4331-bcbe-8964c1c06b41","Type":"ContainerDied","Data":"003fd98c209d40b614bebd1a4ed7f5f1401d682b9848db7a9f1798c17b676ea5"} Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.189220 4989 scope.go:117] "RemoveContainer" containerID="86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.189412 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dk2nj" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.199934 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="806a400ee3481962ca8d84ae2c8fcf07bc8cb4f684fe9dff9bb341ef1146b25a" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.200064 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.251783 4989 scope.go:117] "RemoveContainer" containerID="ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.252943 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.259432 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dk2nj"] Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.283037 4989 scope.go:117] "RemoveContainer" containerID="77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.341223 4989 scope.go:117] "RemoveContainer" containerID="86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.342262 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143\": container with ID starting with 86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143 not found: ID does not exist" containerID="86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.342325 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143"} err="failed to get container status \"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143\": rpc error: code = NotFound desc = could not find container \"86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143\": container with ID starting with 86851c29b211321e775083abe046f803fdcbe87dfb7410afb41174335bc9e143 not found: ID does not exist" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.342369 4989 scope.go:117] "RemoveContainer" containerID="ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.343878 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12\": container with ID starting with ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12 not found: ID does not exist" containerID="ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.343916 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12"} err="failed to get container status \"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12\": rpc error: code = NotFound desc = could not find container \"ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12\": container with ID starting with ea2047d3d9802039a53b77b7acf9b6cd69328d934aa4c1267a8879620db1ba12 not found: ID does not exist" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.343942 4989 scope.go:117] "RemoveContainer" containerID="77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.344246 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575\": container with ID starting with 77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575 not found: ID does not exist" containerID="77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.344295 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575"} err="failed to get container status \"77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575\": rpc error: code = NotFound desc = could not find container \"77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575\": container with ID starting with 77f7eda3cf7c674878e19d6a0dee2f30cf8ce9e89c5aae19ad78e101241a1575 not found: ID does not exist" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.382707 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.383306 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c715de11-638e-41dc-8df9-adfb4f28acbb" containerName="mariadb-client-1" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383333 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c715de11-638e-41dc-8df9-adfb4f28acbb" containerName="mariadb-client-1" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.383381 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="registry-server" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383394 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="registry-server" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.383419 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="extract-content" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383431 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="extract-content" Oct 06 10:01:43 crc kubenswrapper[4989]: E1006 10:01:43.383454 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="extract-utilities" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383466 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="extract-utilities" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383786 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" containerName="registry-server" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.383831 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c715de11-638e-41dc-8df9-adfb4f28acbb" containerName="mariadb-client-1" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.384576 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.392836 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.395384 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.460051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lvrs\" (UniqueName: \"kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs\") pod \"mariadb-client-4-default\" (UID: \"b24d3425-5ef9-4d57-92cd-962506565bd6\") " pod="openstack/mariadb-client-4-default" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.561703 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lvrs\" (UniqueName: \"kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs\") pod \"mariadb-client-4-default\" (UID: \"b24d3425-5ef9-4d57-92cd-962506565bd6\") " pod="openstack/mariadb-client-4-default" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.581766 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lvrs\" (UniqueName: \"kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs\") pod \"mariadb-client-4-default\" (UID: \"b24d3425-5ef9-4d57-92cd-962506565bd6\") " pod="openstack/mariadb-client-4-default" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.712173 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.958511 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a1cad76-c7a2-4331-bcbe-8964c1c06b41" path="/var/lib/kubelet/pods/0a1cad76-c7a2-4331-bcbe-8964c1c06b41/volumes" Oct 06 10:01:43 crc kubenswrapper[4989]: I1006 10:01:43.961089 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c715de11-638e-41dc-8df9-adfb4f28acbb" path="/var/lib/kubelet/pods/c715de11-638e-41dc-8df9-adfb4f28acbb/volumes" Oct 06 10:01:44 crc kubenswrapper[4989]: I1006 10:01:44.098993 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 06 10:01:44 crc kubenswrapper[4989]: I1006 10:01:44.211438 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"b24d3425-5ef9-4d57-92cd-962506565bd6","Type":"ContainerStarted","Data":"95a0f58c22fa45b14514a211a27e995995bdcbfc64d5754c9ba69aabaab77548"} Oct 06 10:01:45 crc kubenswrapper[4989]: I1006 10:01:45.223847 4989 generic.go:334] "Generic (PLEG): container finished" podID="b24d3425-5ef9-4d57-92cd-962506565bd6" containerID="997848024c9b9185ee273592d18e6d5df21125353e0ee828e3bf43b753b11808" exitCode=0 Oct 06 10:01:45 crc kubenswrapper[4989]: I1006 10:01:45.223917 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"b24d3425-5ef9-4d57-92cd-962506565bd6","Type":"ContainerDied","Data":"997848024c9b9185ee273592d18e6d5df21125353e0ee828e3bf43b753b11808"} Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.626763 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.647355 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_b24d3425-5ef9-4d57-92cd-962506565bd6/mariadb-client-4-default/0.log" Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.673181 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.685125 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.714827 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lvrs\" (UniqueName: \"kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs\") pod \"b24d3425-5ef9-4d57-92cd-962506565bd6\" (UID: \"b24d3425-5ef9-4d57-92cd-962506565bd6\") " Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.725448 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs" (OuterVolumeSpecName: "kube-api-access-9lvrs") pod "b24d3425-5ef9-4d57-92cd-962506565bd6" (UID: "b24d3425-5ef9-4d57-92cd-962506565bd6"). InnerVolumeSpecName "kube-api-access-9lvrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:46 crc kubenswrapper[4989]: I1006 10:01:46.816228 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lvrs\" (UniqueName: \"kubernetes.io/projected/b24d3425-5ef9-4d57-92cd-962506565bd6-kube-api-access-9lvrs\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:47 crc kubenswrapper[4989]: I1006 10:01:47.250507 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a0f58c22fa45b14514a211a27e995995bdcbfc64d5754c9ba69aabaab77548" Oct 06 10:01:47 crc kubenswrapper[4989]: I1006 10:01:47.250603 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 06 10:01:47 crc kubenswrapper[4989]: I1006 10:01:47.952951 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24d3425-5ef9-4d57-92cd-962506565bd6" path="/var/lib/kubelet/pods/b24d3425-5ef9-4d57-92cd-962506565bd6/volumes" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.732768 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 06 10:01:51 crc kubenswrapper[4989]: E1006 10:01:51.733516 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24d3425-5ef9-4d57-92cd-962506565bd6" containerName="mariadb-client-4-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.733536 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24d3425-5ef9-4d57-92cd-962506565bd6" containerName="mariadb-client-4-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.733729 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24d3425-5ef9-4d57-92cd-962506565bd6" containerName="mariadb-client-4-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.734329 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.736676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.753270 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.805893 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrdd5\" (UniqueName: \"kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5\") pod \"mariadb-client-5-default\" (UID: \"0c4d7c38-c284-4d9b-853d-1bf6022fe43d\") " pod="openstack/mariadb-client-5-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.907833 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrdd5\" (UniqueName: \"kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5\") pod \"mariadb-client-5-default\" (UID: \"0c4d7c38-c284-4d9b-853d-1bf6022fe43d\") " pod="openstack/mariadb-client-5-default" Oct 06 10:01:51 crc kubenswrapper[4989]: I1006 10:01:51.933015 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrdd5\" (UniqueName: \"kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5\") pod \"mariadb-client-5-default\" (UID: \"0c4d7c38-c284-4d9b-853d-1bf6022fe43d\") " pod="openstack/mariadb-client-5-default" Oct 06 10:01:52 crc kubenswrapper[4989]: I1006 10:01:52.054600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 06 10:01:52 crc kubenswrapper[4989]: I1006 10:01:52.671288 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 06 10:01:52 crc kubenswrapper[4989]: W1006 10:01:52.685686 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c4d7c38_c284_4d9b_853d_1bf6022fe43d.slice/crio-379f5d6ec5c5efa220d5181477fbca42f3d4fca2b394857d051d68626c0c4d7e WatchSource:0}: Error finding container 379f5d6ec5c5efa220d5181477fbca42f3d4fca2b394857d051d68626c0c4d7e: Status 404 returned error can't find the container with id 379f5d6ec5c5efa220d5181477fbca42f3d4fca2b394857d051d68626c0c4d7e Oct 06 10:01:53 crc kubenswrapper[4989]: I1006 10:01:53.305473 4989 generic.go:334] "Generic (PLEG): container finished" podID="0c4d7c38-c284-4d9b-853d-1bf6022fe43d" containerID="d017868364c6651a0b554c3f50fcbb8eff0538a85118f7409530e13d506b71e4" exitCode=0 Oct 06 10:01:53 crc kubenswrapper[4989]: I1006 10:01:53.305537 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"0c4d7c38-c284-4d9b-853d-1bf6022fe43d","Type":"ContainerDied","Data":"d017868364c6651a0b554c3f50fcbb8eff0538a85118f7409530e13d506b71e4"} Oct 06 10:01:53 crc kubenswrapper[4989]: I1006 10:01:53.305599 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"0c4d7c38-c284-4d9b-853d-1bf6022fe43d","Type":"ContainerStarted","Data":"379f5d6ec5c5efa220d5181477fbca42f3d4fca2b394857d051d68626c0c4d7e"} Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.762505 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.785383 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_0c4d7c38-c284-4d9b-853d-1bf6022fe43d/mariadb-client-5-default/0.log" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.817412 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.823582 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.865642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrdd5\" (UniqueName: \"kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5\") pod \"0c4d7c38-c284-4d9b-853d-1bf6022fe43d\" (UID: \"0c4d7c38-c284-4d9b-853d-1bf6022fe43d\") " Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.898121 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5" (OuterVolumeSpecName: "kube-api-access-nrdd5") pod "0c4d7c38-c284-4d9b-853d-1bf6022fe43d" (UID: "0c4d7c38-c284-4d9b-853d-1bf6022fe43d"). InnerVolumeSpecName "kube-api-access-nrdd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.968314 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrdd5\" (UniqueName: \"kubernetes.io/projected/0c4d7c38-c284-4d9b-853d-1bf6022fe43d-kube-api-access-nrdd5\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.994900 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 06 10:01:54 crc kubenswrapper[4989]: E1006 10:01:54.995419 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4d7c38-c284-4d9b-853d-1bf6022fe43d" containerName="mariadb-client-5-default" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.995442 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4d7c38-c284-4d9b-853d-1bf6022fe43d" containerName="mariadb-client-5-default" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.995619 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4d7c38-c284-4d9b-853d-1bf6022fe43d" containerName="mariadb-client-5-default" Oct 06 10:01:54 crc kubenswrapper[4989]: I1006 10:01:54.996417 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.002436 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.068942 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89rp5\" (UniqueName: \"kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5\") pod \"mariadb-client-6-default\" (UID: \"0d3d2123-1bb6-447d-bc86-41ff4570a8ea\") " pod="openstack/mariadb-client-6-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.170774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89rp5\" (UniqueName: \"kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5\") pod \"mariadb-client-6-default\" (UID: \"0d3d2123-1bb6-447d-bc86-41ff4570a8ea\") " pod="openstack/mariadb-client-6-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.191837 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89rp5\" (UniqueName: \"kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5\") pod \"mariadb-client-6-default\" (UID: \"0d3d2123-1bb6-447d-bc86-41ff4570a8ea\") " pod="openstack/mariadb-client-6-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.324222 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="379f5d6ec5c5efa220d5181477fbca42f3d4fca2b394857d051d68626c0c4d7e" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.324302 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.376967 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.909258 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 06 10:01:55 crc kubenswrapper[4989]: I1006 10:01:55.946573 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4d7c38-c284-4d9b-853d-1bf6022fe43d" path="/var/lib/kubelet/pods/0c4d7c38-c284-4d9b-853d-1bf6022fe43d/volumes" Oct 06 10:01:56 crc kubenswrapper[4989]: I1006 10:01:56.335896 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0d3d2123-1bb6-447d-bc86-41ff4570a8ea","Type":"ContainerStarted","Data":"8ea9658e6471defa08591c862404ca7ebfe7d45bf4349daa637e2559c24e8c21"} Oct 06 10:01:56 crc kubenswrapper[4989]: I1006 10:01:56.336263 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0d3d2123-1bb6-447d-bc86-41ff4570a8ea","Type":"ContainerStarted","Data":"12f578e1df42b7abeb5307b7598448a1cf538d2a958b3d1db2778c727a762b72"} Oct 06 10:01:57 crc kubenswrapper[4989]: I1006 10:01:57.349590 4989 generic.go:334] "Generic (PLEG): container finished" podID="0d3d2123-1bb6-447d-bc86-41ff4570a8ea" containerID="8ea9658e6471defa08591c862404ca7ebfe7d45bf4349daa637e2559c24e8c21" exitCode=0 Oct 06 10:01:57 crc kubenswrapper[4989]: I1006 10:01:57.349630 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0d3d2123-1bb6-447d-bc86-41ff4570a8ea","Type":"ContainerDied","Data":"8ea9658e6471defa08591c862404ca7ebfe7d45bf4349daa637e2559c24e8c21"} Oct 06 10:01:58 crc kubenswrapper[4989]: I1006 10:01:58.886625 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 06 10:01:58 crc kubenswrapper[4989]: I1006 10:01:58.938166 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 06 10:01:58 crc kubenswrapper[4989]: I1006 10:01:58.949074 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.058486 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89rp5\" (UniqueName: \"kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5\") pod \"0d3d2123-1bb6-447d-bc86-41ff4570a8ea\" (UID: \"0d3d2123-1bb6-447d-bc86-41ff4570a8ea\") " Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.065040 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5" (OuterVolumeSpecName: "kube-api-access-89rp5") pod "0d3d2123-1bb6-447d-bc86-41ff4570a8ea" (UID: "0d3d2123-1bb6-447d-bc86-41ff4570a8ea"). InnerVolumeSpecName "kube-api-access-89rp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.097431 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 06 10:01:59 crc kubenswrapper[4989]: E1006 10:01:59.097927 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3d2123-1bb6-447d-bc86-41ff4570a8ea" containerName="mariadb-client-6-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.097956 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3d2123-1bb6-447d-bc86-41ff4570a8ea" containerName="mariadb-client-6-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.098269 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3d2123-1bb6-447d-bc86-41ff4570a8ea" containerName="mariadb-client-6-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.099467 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.115007 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.160457 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89rp5\" (UniqueName: \"kubernetes.io/projected/0d3d2123-1bb6-447d-bc86-41ff4570a8ea-kube-api-access-89rp5\") on node \"crc\" DevicePath \"\"" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.261827 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drr4d\" (UniqueName: \"kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d\") pod \"mariadb-client-7-default\" (UID: \"42f733bf-c2ef-431c-a41c-65de240d7f8b\") " pod="openstack/mariadb-client-7-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.363361 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drr4d\" (UniqueName: \"kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d\") pod \"mariadb-client-7-default\" (UID: \"42f733bf-c2ef-431c-a41c-65de240d7f8b\") " pod="openstack/mariadb-client-7-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.368817 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12f578e1df42b7abeb5307b7598448a1cf538d2a958b3d1db2778c727a762b72" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.368878 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.382012 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drr4d\" (UniqueName: \"kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d\") pod \"mariadb-client-7-default\" (UID: \"42f733bf-c2ef-431c-a41c-65de240d7f8b\") " pod="openstack/mariadb-client-7-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.445977 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 06 10:01:59 crc kubenswrapper[4989]: I1006 10:01:59.951955 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3d2123-1bb6-447d-bc86-41ff4570a8ea" path="/var/lib/kubelet/pods/0d3d2123-1bb6-447d-bc86-41ff4570a8ea/volumes" Oct 06 10:02:00 crc kubenswrapper[4989]: I1006 10:02:00.026861 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 06 10:02:00 crc kubenswrapper[4989]: W1006 10:02:00.035994 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42f733bf_c2ef_431c_a41c_65de240d7f8b.slice/crio-c2d62e874d03845c0c40e0f856bb186b004a3b876693b9ff741398a05dafe9fd WatchSource:0}: Error finding container c2d62e874d03845c0c40e0f856bb186b004a3b876693b9ff741398a05dafe9fd: Status 404 returned error can't find the container with id c2d62e874d03845c0c40e0f856bb186b004a3b876693b9ff741398a05dafe9fd Oct 06 10:02:00 crc kubenswrapper[4989]: I1006 10:02:00.389641 4989 generic.go:334] "Generic (PLEG): container finished" podID="42f733bf-c2ef-431c-a41c-65de240d7f8b" containerID="c1ac576c81166aecc1caafd707d246cd1cf381748ba8b714806079c6887e020b" exitCode=0 Oct 06 10:02:00 crc kubenswrapper[4989]: I1006 10:02:00.389996 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"42f733bf-c2ef-431c-a41c-65de240d7f8b","Type":"ContainerDied","Data":"c1ac576c81166aecc1caafd707d246cd1cf381748ba8b714806079c6887e020b"} Oct 06 10:02:00 crc kubenswrapper[4989]: I1006 10:02:00.390417 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"42f733bf-c2ef-431c-a41c-65de240d7f8b","Type":"ContainerStarted","Data":"c2d62e874d03845c0c40e0f856bb186b004a3b876693b9ff741398a05dafe9fd"} Oct 06 10:02:01 crc kubenswrapper[4989]: I1006 10:02:01.920558 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 06 10:02:01 crc kubenswrapper[4989]: I1006 10:02:01.941919 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_42f733bf-c2ef-431c-a41c-65de240d7f8b/mariadb-client-7-default/0.log" Oct 06 10:02:01 crc kubenswrapper[4989]: I1006 10:02:01.968421 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 06 10:02:01 crc kubenswrapper[4989]: I1006 10:02:01.975208 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.112743 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drr4d\" (UniqueName: \"kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d\") pod \"42f733bf-c2ef-431c-a41c-65de240d7f8b\" (UID: \"42f733bf-c2ef-431c-a41c-65de240d7f8b\") " Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.123094 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d" (OuterVolumeSpecName: "kube-api-access-drr4d") pod "42f733bf-c2ef-431c-a41c-65de240d7f8b" (UID: "42f733bf-c2ef-431c-a41c-65de240d7f8b"). InnerVolumeSpecName "kube-api-access-drr4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.129298 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 06 10:02:02 crc kubenswrapper[4989]: E1006 10:02:02.129940 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f733bf-c2ef-431c-a41c-65de240d7f8b" containerName="mariadb-client-7-default" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.129973 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f733bf-c2ef-431c-a41c-65de240d7f8b" containerName="mariadb-client-7-default" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.130274 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f733bf-c2ef-431c-a41c-65de240d7f8b" containerName="mariadb-client-7-default" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.131290 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.138113 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.216169 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drr4d\" (UniqueName: \"kubernetes.io/projected/42f733bf-c2ef-431c-a41c-65de240d7f8b-kube-api-access-drr4d\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.318875 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzjdd\" (UniqueName: \"kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd\") pod \"mariadb-client-2\" (UID: \"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f\") " pod="openstack/mariadb-client-2" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.409301 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2d62e874d03845c0c40e0f856bb186b004a3b876693b9ff741398a05dafe9fd" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.409396 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.420065 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzjdd\" (UniqueName: \"kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd\") pod \"mariadb-client-2\" (UID: \"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f\") " pod="openstack/mariadb-client-2" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.448772 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzjdd\" (UniqueName: \"kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd\") pod \"mariadb-client-2\" (UID: \"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f\") " pod="openstack/mariadb-client-2" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.470994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 06 10:02:02 crc kubenswrapper[4989]: I1006 10:02:02.923235 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 06 10:02:03 crc kubenswrapper[4989]: I1006 10:02:03.419117 4989 generic.go:334] "Generic (PLEG): container finished" podID="22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" containerID="87f7a737a3b525eecca89b0c2ac0dbfce46231f8e326b3f9960964556de8cb03" exitCode=0 Oct 06 10:02:03 crc kubenswrapper[4989]: I1006 10:02:03.419230 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f","Type":"ContainerDied","Data":"87f7a737a3b525eecca89b0c2ac0dbfce46231f8e326b3f9960964556de8cb03"} Oct 06 10:02:03 crc kubenswrapper[4989]: I1006 10:02:03.419482 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f","Type":"ContainerStarted","Data":"684e00920410457921accd6f1267ded299d979dc89f06d2704e30291d4ae7e21"} Oct 06 10:02:03 crc kubenswrapper[4989]: I1006 10:02:03.951619 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f733bf-c2ef-431c-a41c-65de240d7f8b" path="/var/lib/kubelet/pods/42f733bf-c2ef-431c-a41c-65de240d7f8b/volumes" Oct 06 10:02:04 crc kubenswrapper[4989]: I1006 10:02:04.932389 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 06 10:02:04 crc kubenswrapper[4989]: I1006 10:02:04.948866 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_22989cc8-7fb5-47f8-bf4e-3f767e04aa7f/mariadb-client-2/0.log" Oct 06 10:02:04 crc kubenswrapper[4989]: I1006 10:02:04.973313 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 06 10:02:04 crc kubenswrapper[4989]: I1006 10:02:04.979694 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.060629 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzjdd\" (UniqueName: \"kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd\") pod \"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f\" (UID: \"22989cc8-7fb5-47f8-bf4e-3f767e04aa7f\") " Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.070396 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd" (OuterVolumeSpecName: "kube-api-access-wzjdd") pod "22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" (UID: "22989cc8-7fb5-47f8-bf4e-3f767e04aa7f"). InnerVolumeSpecName "kube-api-access-wzjdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.162533 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzjdd\" (UniqueName: \"kubernetes.io/projected/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f-kube-api-access-wzjdd\") on node \"crc\" DevicePath \"\"" Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.436171 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="684e00920410457921accd6f1267ded299d979dc89f06d2704e30291d4ae7e21" Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.436268 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 06 10:02:05 crc kubenswrapper[4989]: I1006 10:02:05.950062 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" path="/var/lib/kubelet/pods/22989cc8-7fb5-47f8-bf4e-3f767e04aa7f/volumes" Oct 06 10:02:19 crc kubenswrapper[4989]: I1006 10:02:19.984856 4989 scope.go:117] "RemoveContainer" containerID="e46f6268a8f0cd083838f3377bb458f404b3dc276905773e00620ec1986b6b4a" Oct 06 10:03:03 crc kubenswrapper[4989]: I1006 10:03:03.935055 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:03:03 crc kubenswrapper[4989]: I1006 10:03:03.935785 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:03:33 crc kubenswrapper[4989]: I1006 10:03:33.935970 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:03:33 crc kubenswrapper[4989]: I1006 10:03:33.939063 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:04:03 crc kubenswrapper[4989]: I1006 10:04:03.935995 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:04:03 crc kubenswrapper[4989]: I1006 10:04:03.936688 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:04:03 crc kubenswrapper[4989]: I1006 10:04:03.949864 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:04:03 crc kubenswrapper[4989]: I1006 10:04:03.950865 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:04:03 crc kubenswrapper[4989]: I1006 10:04:03.950965 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b" gracePeriod=600 Oct 06 10:04:04 crc kubenswrapper[4989]: I1006 10:04:04.722420 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b" exitCode=0 Oct 06 10:04:04 crc kubenswrapper[4989]: I1006 10:04:04.722493 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b"} Oct 06 10:04:04 crc kubenswrapper[4989]: I1006 10:04:04.722824 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5"} Oct 06 10:04:04 crc kubenswrapper[4989]: I1006 10:04:04.722858 4989 scope.go:117] "RemoveContainer" containerID="3fcb59c988ae99aed972a114617e8d5e933d57d0b41a5bedb15dec23d8f384ae" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.478151 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 10:06:18 crc kubenswrapper[4989]: E1006 10:06:18.479705 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" containerName="mariadb-client-2" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.479739 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" containerName="mariadb-client-2" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.480122 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="22989cc8-7fb5-47f8-bf4e-3f767e04aa7f" containerName="mariadb-client-2" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.481380 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.485568 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-76mdq" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.490239 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.633477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb4sc\" (UniqueName: \"kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.633832 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.735719 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.735836 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb4sc\" (UniqueName: \"kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.738859 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.738893 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ec741f57cdbe2de573e0254f0c8d9f7bcb8ccddc254a5316eac0147b3e0db755/globalmount\"" pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.761596 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb4sc\" (UniqueName: \"kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.778452 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") pod \"mariadb-copy-data\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " pod="openstack/mariadb-copy-data" Oct 06 10:06:18 crc kubenswrapper[4989]: I1006 10:06:18.811015 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 06 10:06:19 crc kubenswrapper[4989]: I1006 10:06:19.372320 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 10:06:20 crc kubenswrapper[4989]: I1006 10:06:20.158000 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3","Type":"ContainerStarted","Data":"f8c31a9b0f716d1fa7864879dbc8d9457f62b077bf9eafb9939aedc1d96f90e0"} Oct 06 10:06:20 crc kubenswrapper[4989]: I1006 10:06:20.158623 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3","Type":"ContainerStarted","Data":"a007c5219903258b6763b630ab901d11bb0c315ba99c9491548baadfb59dc164"} Oct 06 10:06:20 crc kubenswrapper[4989]: I1006 10:06:20.176310 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.176287553 podStartE2EDuration="3.176287553s" podCreationTimestamp="2025-10-06 10:06:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:06:20.170715232 +0000 UTC m=+5230.960740812" watchObservedRunningTime="2025-10-06 10:06:20.176287553 +0000 UTC m=+5230.966313123" Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.211293 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.212510 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.220091 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.296539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7kdb\" (UniqueName: \"kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb\") pod \"mariadb-client\" (UID: \"aa60356d-8579-4e51-80cf-3721ae6160bd\") " pod="openstack/mariadb-client" Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.398817 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7kdb\" (UniqueName: \"kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb\") pod \"mariadb-client\" (UID: \"aa60356d-8579-4e51-80cf-3721ae6160bd\") " pod="openstack/mariadb-client" Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.421272 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7kdb\" (UniqueName: \"kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb\") pod \"mariadb-client\" (UID: \"aa60356d-8579-4e51-80cf-3721ae6160bd\") " pod="openstack/mariadb-client" Oct 06 10:06:22 crc kubenswrapper[4989]: I1006 10:06:22.545491 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:23 crc kubenswrapper[4989]: I1006 10:06:22.999633 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:23 crc kubenswrapper[4989]: I1006 10:06:23.192509 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"aa60356d-8579-4e51-80cf-3721ae6160bd","Type":"ContainerStarted","Data":"a627e6e1bc389a549cdbe0b6236d7583aad2ecb9c598a7c702f48815b11c4c95"} Oct 06 10:06:23 crc kubenswrapper[4989]: I1006 10:06:23.192590 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"aa60356d-8579-4e51-80cf-3721ae6160bd","Type":"ContainerStarted","Data":"1ce73fbe5fb8d3a6155934670b1a9afd0244a9debd9e4bb86ba30d1d1f0ee544"} Oct 06 10:06:23 crc kubenswrapper[4989]: I1006 10:06:23.208159 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=1.208136453 podStartE2EDuration="1.208136453s" podCreationTimestamp="2025-10-06 10:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:06:23.207500415 +0000 UTC m=+5233.997526035" watchObservedRunningTime="2025-10-06 10:06:23.208136453 +0000 UTC m=+5233.998162033" Oct 06 10:06:23 crc kubenswrapper[4989]: I1006 10:06:23.253057 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_aa60356d-8579-4e51-80cf-3721ae6160bd/mariadb-client/0.log" Oct 06 10:06:24 crc kubenswrapper[4989]: I1006 10:06:24.204561 4989 generic.go:334] "Generic (PLEG): container finished" podID="aa60356d-8579-4e51-80cf-3721ae6160bd" containerID="a627e6e1bc389a549cdbe0b6236d7583aad2ecb9c598a7c702f48815b11c4c95" exitCode=0 Oct 06 10:06:24 crc kubenswrapper[4989]: I1006 10:06:24.204749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"aa60356d-8579-4e51-80cf-3721ae6160bd","Type":"ContainerDied","Data":"a627e6e1bc389a549cdbe0b6236d7583aad2ecb9c598a7c702f48815b11c4c95"} Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.605018 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.649578 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.655675 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7kdb\" (UniqueName: \"kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb\") pod \"aa60356d-8579-4e51-80cf-3721ae6160bd\" (UID: \"aa60356d-8579-4e51-80cf-3721ae6160bd\") " Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.657115 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.669095 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb" (OuterVolumeSpecName: "kube-api-access-w7kdb") pod "aa60356d-8579-4e51-80cf-3721ae6160bd" (UID: "aa60356d-8579-4e51-80cf-3721ae6160bd"). InnerVolumeSpecName "kube-api-access-w7kdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.757987 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7kdb\" (UniqueName: \"kubernetes.io/projected/aa60356d-8579-4e51-80cf-3721ae6160bd-kube-api-access-w7kdb\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.790583 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:25 crc kubenswrapper[4989]: E1006 10:06:25.790951 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa60356d-8579-4e51-80cf-3721ae6160bd" containerName="mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.790968 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa60356d-8579-4e51-80cf-3721ae6160bd" containerName="mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.791137 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa60356d-8579-4e51-80cf-3721ae6160bd" containerName="mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.791638 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.803709 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.861532 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkrbr\" (UniqueName: \"kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr\") pod \"mariadb-client\" (UID: \"95ae6510-4bea-47c0-aa6e-64ddd9c133c9\") " pod="openstack/mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.954996 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa60356d-8579-4e51-80cf-3721ae6160bd" path="/var/lib/kubelet/pods/aa60356d-8579-4e51-80cf-3721ae6160bd/volumes" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.963197 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkrbr\" (UniqueName: \"kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr\") pod \"mariadb-client\" (UID: \"95ae6510-4bea-47c0-aa6e-64ddd9c133c9\") " pod="openstack/mariadb-client" Oct 06 10:06:25 crc kubenswrapper[4989]: I1006 10:06:25.985627 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkrbr\" (UniqueName: \"kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr\") pod \"mariadb-client\" (UID: \"95ae6510-4bea-47c0-aa6e-64ddd9c133c9\") " pod="openstack/mariadb-client" Oct 06 10:06:26 crc kubenswrapper[4989]: I1006 10:06:26.132112 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:26 crc kubenswrapper[4989]: I1006 10:06:26.232021 4989 scope.go:117] "RemoveContainer" containerID="a627e6e1bc389a549cdbe0b6236d7583aad2ecb9c598a7c702f48815b11c4c95" Oct 06 10:06:26 crc kubenswrapper[4989]: I1006 10:06:26.232081 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:26 crc kubenswrapper[4989]: I1006 10:06:26.359384 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:26 crc kubenswrapper[4989]: W1006 10:06:26.360557 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95ae6510_4bea_47c0_aa6e_64ddd9c133c9.slice/crio-61283a7bed10e6ab342b582dc451fa007e2e6769fc7003120cefcb3922898a64 WatchSource:0}: Error finding container 61283a7bed10e6ab342b582dc451fa007e2e6769fc7003120cefcb3922898a64: Status 404 returned error can't find the container with id 61283a7bed10e6ab342b582dc451fa007e2e6769fc7003120cefcb3922898a64 Oct 06 10:06:27 crc kubenswrapper[4989]: I1006 10:06:27.247809 4989 generic.go:334] "Generic (PLEG): container finished" podID="95ae6510-4bea-47c0-aa6e-64ddd9c133c9" containerID="8d8fb23139b64792e555c273cae30b610a82d5e23253fa4124f66c1e9d64794a" exitCode=0 Oct 06 10:06:27 crc kubenswrapper[4989]: I1006 10:06:27.247907 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"95ae6510-4bea-47c0-aa6e-64ddd9c133c9","Type":"ContainerDied","Data":"8d8fb23139b64792e555c273cae30b610a82d5e23253fa4124f66c1e9d64794a"} Oct 06 10:06:27 crc kubenswrapper[4989]: I1006 10:06:27.248250 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"95ae6510-4bea-47c0-aa6e-64ddd9c133c9","Type":"ContainerStarted","Data":"61283a7bed10e6ab342b582dc451fa007e2e6769fc7003120cefcb3922898a64"} Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.567241 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.585980 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_95ae6510-4bea-47c0-aa6e-64ddd9c133c9/mariadb-client/0.log" Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.607501 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkrbr\" (UniqueName: \"kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr\") pod \"95ae6510-4bea-47c0-aa6e-64ddd9c133c9\" (UID: \"95ae6510-4bea-47c0-aa6e-64ddd9c133c9\") " Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.613910 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr" (OuterVolumeSpecName: "kube-api-access-pkrbr") pod "95ae6510-4bea-47c0-aa6e-64ddd9c133c9" (UID: "95ae6510-4bea-47c0-aa6e-64ddd9c133c9"). InnerVolumeSpecName "kube-api-access-pkrbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.616766 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.623751 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 06 10:06:28 crc kubenswrapper[4989]: I1006 10:06:28.709988 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkrbr\" (UniqueName: \"kubernetes.io/projected/95ae6510-4bea-47c0-aa6e-64ddd9c133c9-kube-api-access-pkrbr\") on node \"crc\" DevicePath \"\"" Oct 06 10:06:29 crc kubenswrapper[4989]: I1006 10:06:29.274344 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61283a7bed10e6ab342b582dc451fa007e2e6769fc7003120cefcb3922898a64" Oct 06 10:06:29 crc kubenswrapper[4989]: I1006 10:06:29.274995 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 06 10:06:29 crc kubenswrapper[4989]: I1006 10:06:29.968727 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ae6510-4bea-47c0-aa6e-64ddd9c133c9" path="/var/lib/kubelet/pods/95ae6510-4bea-47c0-aa6e-64ddd9c133c9/volumes" Oct 06 10:06:33 crc kubenswrapper[4989]: I1006 10:06:33.935607 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:06:33 crc kubenswrapper[4989]: I1006 10:06:33.936007 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.564596 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:06:50 crc kubenswrapper[4989]: E1006 10:06:50.566379 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ae6510-4bea-47c0-aa6e-64ddd9c133c9" containerName="mariadb-client" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.566403 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ae6510-4bea-47c0-aa6e-64ddd9c133c9" containerName="mariadb-client" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.566750 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ae6510-4bea-47c0-aa6e-64ddd9c133c9" containerName="mariadb-client" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.570026 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.579278 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.672132 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrdjb\" (UniqueName: \"kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.672213 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.672292 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.773928 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrdjb\" (UniqueName: \"kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.774008 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.774044 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.774679 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.774715 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.800544 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrdjb\" (UniqueName: \"kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb\") pod \"redhat-marketplace-qkn99\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:50 crc kubenswrapper[4989]: I1006 10:06:50.895018 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.157410 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.161334 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.166134 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.280740 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.280803 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.280863 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7468k\" (UniqueName: \"kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.336309 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.382611 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.383289 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7468k\" (UniqueName: \"kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.383706 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.384122 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.384217 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.407256 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7468k\" (UniqueName: \"kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k\") pod \"community-operators-dwpzb\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.482862 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerStarted","Data":"2ee2bd197c7725e0f2d89f500cc5d39e695125ba09b15c6ce900e010f817a7be"} Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.484908 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:06:51 crc kubenswrapper[4989]: I1006 10:06:51.959276 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:06:51 crc kubenswrapper[4989]: W1006 10:06:51.972193 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e8cba0b_c738_4464_86f1_7c08badf9e85.slice/crio-903100b870e51daa7e12a12e0db42a97a9236aaeb522b95ce548aa68ea57632f WatchSource:0}: Error finding container 903100b870e51daa7e12a12e0db42a97a9236aaeb522b95ce548aa68ea57632f: Status 404 returned error can't find the container with id 903100b870e51daa7e12a12e0db42a97a9236aaeb522b95ce548aa68ea57632f Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.496068 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerID="759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96" exitCode=0 Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.496127 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerDied","Data":"759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96"} Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.496639 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerStarted","Data":"903100b870e51daa7e12a12e0db42a97a9236aaeb522b95ce548aa68ea57632f"} Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.500095 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.500961 4989 generic.go:334] "Generic (PLEG): container finished" podID="53dec518-247a-4933-a9f5-fbe610906805" containerID="39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681" exitCode=0 Oct 06 10:06:52 crc kubenswrapper[4989]: I1006 10:06:52.501007 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerDied","Data":"39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681"} Oct 06 10:06:53 crc kubenswrapper[4989]: I1006 10:06:53.510147 4989 generic.go:334] "Generic (PLEG): container finished" podID="53dec518-247a-4933-a9f5-fbe610906805" containerID="78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1" exitCode=0 Oct 06 10:06:53 crc kubenswrapper[4989]: I1006 10:06:53.510199 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerDied","Data":"78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1"} Oct 06 10:06:54 crc kubenswrapper[4989]: I1006 10:06:54.521526 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerID="e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a" exitCode=0 Oct 06 10:06:54 crc kubenswrapper[4989]: I1006 10:06:54.521803 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerDied","Data":"e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a"} Oct 06 10:06:54 crc kubenswrapper[4989]: I1006 10:06:54.526212 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerStarted","Data":"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1"} Oct 06 10:06:55 crc kubenswrapper[4989]: I1006 10:06:55.538744 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerStarted","Data":"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa"} Oct 06 10:06:55 crc kubenswrapper[4989]: I1006 10:06:55.560215 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qkn99" podStartSLOduration=3.87455454 podStartE2EDuration="5.560197878s" podCreationTimestamp="2025-10-06 10:06:50 +0000 UTC" firstStartedPulling="2025-10-06 10:06:52.503372718 +0000 UTC m=+5263.293398318" lastFinishedPulling="2025-10-06 10:06:54.189016056 +0000 UTC m=+5264.979041656" observedRunningTime="2025-10-06 10:06:54.567511307 +0000 UTC m=+5265.357536907" watchObservedRunningTime="2025-10-06 10:06:55.560197878 +0000 UTC m=+5266.350223458" Oct 06 10:06:55 crc kubenswrapper[4989]: I1006 10:06:55.564875 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dwpzb" podStartSLOduration=2.134205267 podStartE2EDuration="4.564864952s" podCreationTimestamp="2025-10-06 10:06:51 +0000 UTC" firstStartedPulling="2025-10-06 10:06:52.499448635 +0000 UTC m=+5263.289474225" lastFinishedPulling="2025-10-06 10:06:54.93010833 +0000 UTC m=+5265.720133910" observedRunningTime="2025-10-06 10:06:55.564150992 +0000 UTC m=+5266.354176612" watchObservedRunningTime="2025-10-06 10:06:55.564864952 +0000 UTC m=+5266.354890532" Oct 06 10:07:00 crc kubenswrapper[4989]: I1006 10:07:00.895526 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:00 crc kubenswrapper[4989]: I1006 10:07:00.896409 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:00 crc kubenswrapper[4989]: I1006 10:07:00.977341 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:01 crc kubenswrapper[4989]: I1006 10:07:01.486012 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:01 crc kubenswrapper[4989]: I1006 10:07:01.486103 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:01 crc kubenswrapper[4989]: I1006 10:07:01.535574 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:01 crc kubenswrapper[4989]: I1006 10:07:01.679967 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:01 crc kubenswrapper[4989]: I1006 10:07:01.685291 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:02 crc kubenswrapper[4989]: I1006 10:07:02.621440 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:07:03 crc kubenswrapper[4989]: I1006 10:07:03.631568 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dwpzb" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="registry-server" containerID="cri-o://03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa" gracePeriod=2 Oct 06 10:07:03 crc kubenswrapper[4989]: I1006 10:07:03.935764 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:07:03 crc kubenswrapper[4989]: I1006 10:07:03.936286 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.038948 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.039289 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qkn99" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="registry-server" containerID="cri-o://9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1" gracePeriod=2 Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.195859 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.294369 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities\") pod \"4e8cba0b-c738-4464-86f1-7c08badf9e85\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.294427 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7468k\" (UniqueName: \"kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k\") pod \"4e8cba0b-c738-4464-86f1-7c08badf9e85\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.294486 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content\") pod \"4e8cba0b-c738-4464-86f1-7c08badf9e85\" (UID: \"4e8cba0b-c738-4464-86f1-7c08badf9e85\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.295454 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities" (OuterVolumeSpecName: "utilities") pod "4e8cba0b-c738-4464-86f1-7c08badf9e85" (UID: "4e8cba0b-c738-4464-86f1-7c08badf9e85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.306794 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k" (OuterVolumeSpecName: "kube-api-access-7468k") pod "4e8cba0b-c738-4464-86f1-7c08badf9e85" (UID: "4e8cba0b-c738-4464-86f1-7c08badf9e85"). InnerVolumeSpecName "kube-api-access-7468k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.348896 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e8cba0b-c738-4464-86f1-7c08badf9e85" (UID: "4e8cba0b-c738-4464-86f1-7c08badf9e85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.395934 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.395983 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7468k\" (UniqueName: \"kubernetes.io/projected/4e8cba0b-c738-4464-86f1-7c08badf9e85-kube-api-access-7468k\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.395999 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e8cba0b-c738-4464-86f1-7c08badf9e85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.445857 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.598284 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrdjb\" (UniqueName: \"kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb\") pod \"53dec518-247a-4933-a9f5-fbe610906805\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.598360 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content\") pod \"53dec518-247a-4933-a9f5-fbe610906805\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.598448 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities\") pod \"53dec518-247a-4933-a9f5-fbe610906805\" (UID: \"53dec518-247a-4933-a9f5-fbe610906805\") " Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.599382 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities" (OuterVolumeSpecName: "utilities") pod "53dec518-247a-4933-a9f5-fbe610906805" (UID: "53dec518-247a-4933-a9f5-fbe610906805"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.601489 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb" (OuterVolumeSpecName: "kube-api-access-lrdjb") pod "53dec518-247a-4933-a9f5-fbe610906805" (UID: "53dec518-247a-4933-a9f5-fbe610906805"). InnerVolumeSpecName "kube-api-access-lrdjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.611879 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53dec518-247a-4933-a9f5-fbe610906805" (UID: "53dec518-247a-4933-a9f5-fbe610906805"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.645281 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerID="03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa" exitCode=0 Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.645321 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerDied","Data":"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa"} Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.645358 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dwpzb" event={"ID":"4e8cba0b-c738-4464-86f1-7c08badf9e85","Type":"ContainerDied","Data":"903100b870e51daa7e12a12e0db42a97a9236aaeb522b95ce548aa68ea57632f"} Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.645379 4989 scope.go:117] "RemoveContainer" containerID="03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.645449 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dwpzb" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.649635 4989 generic.go:334] "Generic (PLEG): container finished" podID="53dec518-247a-4933-a9f5-fbe610906805" containerID="9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1" exitCode=0 Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.649690 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerDied","Data":"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1"} Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.649719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qkn99" event={"ID":"53dec518-247a-4933-a9f5-fbe610906805","Type":"ContainerDied","Data":"2ee2bd197c7725e0f2d89f500cc5d39e695125ba09b15c6ce900e010f817a7be"} Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.649767 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qkn99" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.669988 4989 scope.go:117] "RemoveContainer" containerID="e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.693335 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.701576 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.701692 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrdjb\" (UniqueName: \"kubernetes.io/projected/53dec518-247a-4933-a9f5-fbe610906805-kube-api-access-lrdjb\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.701710 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53dec518-247a-4933-a9f5-fbe610906805-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.709368 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dwpzb"] Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.717339 4989 scope.go:117] "RemoveContainer" containerID="759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.720104 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.729190 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qkn99"] Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.734471 4989 scope.go:117] "RemoveContainer" containerID="03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.735144 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa\": container with ID starting with 03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa not found: ID does not exist" containerID="03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.735196 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa"} err="failed to get container status \"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa\": rpc error: code = NotFound desc = could not find container \"03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa\": container with ID starting with 03a6fc35446a8a0ed7fdbb943e0c31e22458936e6e5b0bb9a1e86897e39741fa not found: ID does not exist" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.735230 4989 scope.go:117] "RemoveContainer" containerID="e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.735678 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a\": container with ID starting with e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a not found: ID does not exist" containerID="e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.735718 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a"} err="failed to get container status \"e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a\": rpc error: code = NotFound desc = could not find container \"e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a\": container with ID starting with e4213ffeb882008d5319bc5e3ded150038f4f0706c4acdb781385a412488644a not found: ID does not exist" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.735742 4989 scope.go:117] "RemoveContainer" containerID="759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.736039 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96\": container with ID starting with 759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96 not found: ID does not exist" containerID="759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.736128 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96"} err="failed to get container status \"759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96\": rpc error: code = NotFound desc = could not find container \"759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96\": container with ID starting with 759978fc8d614aaacb96cd8ca6f2521c4350d94710ac2b0c4598a069d77c4c96 not found: ID does not exist" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.736210 4989 scope.go:117] "RemoveContainer" containerID="9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.766594 4989 scope.go:117] "RemoveContainer" containerID="78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.821601 4989 scope.go:117] "RemoveContainer" containerID="39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.850249 4989 scope.go:117] "RemoveContainer" containerID="9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.853765 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1\": container with ID starting with 9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1 not found: ID does not exist" containerID="9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.853796 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1"} err="failed to get container status \"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1\": rpc error: code = NotFound desc = could not find container \"9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1\": container with ID starting with 9f0b4d47761626f040d063749189802689a702547ff1588cce633407fc8850b1 not found: ID does not exist" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.853817 4989 scope.go:117] "RemoveContainer" containerID="78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.875782 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1\": container with ID starting with 78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1 not found: ID does not exist" containerID="78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.875838 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1"} err="failed to get container status \"78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1\": rpc error: code = NotFound desc = could not find container \"78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1\": container with ID starting with 78a48b013900186eb5e6eb749a00f6c743f475210935dd6f347a3627ef69a0f1 not found: ID does not exist" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.875872 4989 scope.go:117] "RemoveContainer" containerID="39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681" Oct 06 10:07:04 crc kubenswrapper[4989]: E1006 10:07:04.879812 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681\": container with ID starting with 39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681 not found: ID does not exist" containerID="39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681" Oct 06 10:07:04 crc kubenswrapper[4989]: I1006 10:07:04.879845 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681"} err="failed to get container status \"39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681\": rpc error: code = NotFound desc = could not find container \"39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681\": container with ID starting with 39f4ed33d1ff63c62518efcef2735327d719ae5313288a44d136d720915c7681 not found: ID does not exist" Oct 06 10:07:05 crc kubenswrapper[4989]: I1006 10:07:05.948401 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" path="/var/lib/kubelet/pods/4e8cba0b-c738-4464-86f1-7c08badf9e85/volumes" Oct 06 10:07:05 crc kubenswrapper[4989]: I1006 10:07:05.949602 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53dec518-247a-4933-a9f5-fbe610906805" path="/var/lib/kubelet/pods/53dec518-247a-4933-a9f5-fbe610906805/volumes" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.196473 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197349 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="extract-content" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197373 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="extract-content" Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197404 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197420 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197461 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="extract-content" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197472 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="extract-content" Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197501 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197513 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197534 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="extract-utilities" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197546 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="extract-utilities" Oct 06 10:07:10 crc kubenswrapper[4989]: E1006 10:07:10.197563 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="extract-utilities" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197575 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="extract-utilities" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197895 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="53dec518-247a-4933-a9f5-fbe610906805" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.197930 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8cba0b-c738-4464-86f1-7c08badf9e85" containerName="registry-server" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.199337 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.202327 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.203032 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.206398 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-2kn45" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.209899 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.211997 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.227422 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.229855 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.234566 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.241556 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.272281 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295329 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/c049c234-54cd-4801-9e73-41c5cb8d1404-kube-api-access-x42dz\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295382 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c049c234-54cd-4801-9e73-41c5cb8d1404-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295440 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295472 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c049c234-54cd-4801-9e73-41c5cb8d1404-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295630 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-config\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.295728 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69t2t\" (UniqueName: \"kubernetes.io/projected/0dabb4ae-66bb-4073-af87-e8a398b22b1a-kube-api-access-69t2t\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296691 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296748 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dabb4ae-66bb-4073-af87-e8a398b22b1a-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296840 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-config\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dabb4ae-66bb-4073-af87-e8a398b22b1a-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296908 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.296938 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.297024 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.297079 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.297114 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsc2r\" (UniqueName: \"kubernetes.io/projected/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-kube-api-access-dsc2r\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.297166 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.385117 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.386470 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398582 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398617 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398781 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398842 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dabb4ae-66bb-4073-af87-e8a398b22b1a-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398870 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-config\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dabb4ae-66bb-4073-af87-e8a398b22b1a-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398907 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398924 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398957 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.398986 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399015 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsc2r\" (UniqueName: \"kubernetes.io/projected/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-kube-api-access-dsc2r\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399040 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399071 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/c049c234-54cd-4801-9e73-41c5cb8d1404-kube-api-access-x42dz\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399097 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c049c234-54cd-4801-9e73-41c5cb8d1404-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399155 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c049c234-54cd-4801-9e73-41c5cb8d1404-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399215 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-config\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399234 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399255 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69t2t\" (UniqueName: \"kubernetes.io/projected/0dabb4ae-66bb-4073-af87-e8a398b22b1a-kube-api-access-69t2t\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399273 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399402 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399819 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0dabb4ae-66bb-4073-af87-e8a398b22b1a-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.399996 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jcxnd" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.400824 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.400965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c049c234-54cd-4801-9e73-41c5cb8d1404-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.401428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dabb4ae-66bb-4073-af87-e8a398b22b1a-config\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.401577 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-config\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.402152 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.407576 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.402253 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c049c234-54cd-4801-9e73-41c5cb8d1404-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.408622 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c049c234-54cd-4801-9e73-41c5cb8d1404-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.411919 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.411981 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4fd7bcd5b05bb6f99c0179b281ea1171dd0fe77a0db6ca1d003c7be3e16b37c8/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.412533 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.412586 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1ef04b2325d69ceb29b8f68d19ebd4102df1e199276d7e95a6e3e012b1f6ec1c/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.412772 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.412825 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7986664cad77aa2651319f10dcbefdc29fb7f89d4bc181460a636ce5b870bbb5/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.414935 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dabb4ae-66bb-4073-af87-e8a398b22b1a-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.418834 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.427936 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.429907 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.430869 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.441207 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69t2t\" (UniqueName: \"kubernetes.io/projected/0dabb4ae-66bb-4073-af87-e8a398b22b1a-kube-api-access-69t2t\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.441891 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.444049 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.449733 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.450466 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/c049c234-54cd-4801-9e73-41c5cb8d1404-kube-api-access-x42dz\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.461949 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.469484 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsc2r\" (UniqueName: \"kubernetes.io/projected/af55f2bd-22b6-4604-9d04-03dbe5aaf3a6-kube-api-access-dsc2r\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500125 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500181 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6r2r\" (UniqueName: \"kubernetes.io/projected/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-kube-api-access-c6r2r\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500299 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.500333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-74e61053-8d25-4a15-a3f2-765438c55925\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74e61053-8d25-4a15-a3f2-765438c55925\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.509777 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-adc0e574-d29b-4d6d-b297-50159a8bd1eb\") pod \"ovsdbserver-nb-1\" (UID: \"c049c234-54cd-4801-9e73-41c5cb8d1404\") " pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.510978 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-758f8709-3529-4ed5-93cb-8d8d439fbd38\") pod \"ovsdbserver-nb-2\" (UID: \"0dabb4ae-66bb-4073-af87-e8a398b22b1a\") " pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.531222 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fa118f44-fbf6-4d73-9c48-2ad654a1fc59\") pod \"ovsdbserver-nb-0\" (UID: \"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6\") " pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.549106 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.577815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.592935 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601383 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601464 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-74e61053-8d25-4a15-a3f2-765438c55925\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74e61053-8d25-4a15-a3f2-765438c55925\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601514 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601550 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601583 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3fa8a8-b876-485e-954f-3ed1c711bd61-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601614 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82tcf\" (UniqueName: \"kubernetes.io/projected/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-kube-api-access-82tcf\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601647 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601688 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6r2r\" (UniqueName: \"kubernetes.io/projected/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-kube-api-access-c6r2r\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601740 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601802 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601828 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-config\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601857 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601878 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpdxf\" (UniqueName: \"kubernetes.io/projected/fb3fa8a8-b876-485e-954f-3ed1c711bd61-kube-api-access-dpdxf\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601914 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb3fa8a8-b876-485e-954f-3ed1c711bd61-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601948 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.601974 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.602816 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.604464 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.604669 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.605974 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.606004 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-74e61053-8d25-4a15-a3f2-765438c55925\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74e61053-8d25-4a15-a3f2-765438c55925\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f6665ee2189d51ab35a0d3d10f9b9116d7b660052aa18a096ea2fd9bec99d06b/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.611129 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.622392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6r2r\" (UniqueName: \"kubernetes.io/projected/27ebef9a-25d5-437a-bd77-5b8ccd0110ba-kube-api-access-c6r2r\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.664218 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-74e61053-8d25-4a15-a3f2-765438c55925\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74e61053-8d25-4a15-a3f2-765438c55925\") pod \"ovsdbserver-sb-0\" (UID: \"27ebef9a-25d5-437a-bd77-5b8ccd0110ba\") " pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.702859 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3fa8a8-b876-485e-954f-3ed1c711bd61-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.702906 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82tcf\" (UniqueName: \"kubernetes.io/projected/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-kube-api-access-82tcf\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.702933 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.702990 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703024 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-config\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703044 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703059 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpdxf\" (UniqueName: \"kubernetes.io/projected/fb3fa8a8-b876-485e-954f-3ed1c711bd61-kube-api-access-dpdxf\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703082 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb3fa8a8-b876-485e-954f-3ed1c711bd61-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703106 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703173 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703225 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.703252 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.704275 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb3fa8a8-b876-485e-954f-3ed1c711bd61-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.704312 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.704744 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.705015 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb3fa8a8-b876-485e-954f-3ed1c711bd61-config\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.705405 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-config\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.705953 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.708101 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.708131 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/88a9c6c7286fc6babf4424e23fb7bcc0984507bf8b28d8484914ef597d5695b4/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.708348 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.708424 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf5e535ce3e3bad0bfc5af83af40b9e0a403ede001f0ddad4a737e7ead481308/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.720939 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3fa8a8-b876-485e-954f-3ed1c711bd61-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.721246 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.721583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpdxf\" (UniqueName: \"kubernetes.io/projected/fb3fa8a8-b876-485e-954f-3ed1c711bd61-kube-api-access-dpdxf\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.722955 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82tcf\" (UniqueName: \"kubernetes.io/projected/bb75f7eb-e857-4e0c-8900-bd9fea672fb5-kube-api-access-82tcf\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.736422 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-443719c6-dbe6-459a-a7ba-8d34a578a804\") pod \"ovsdbserver-sb-2\" (UID: \"fb3fa8a8-b876-485e-954f-3ed1c711bd61\") " pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.737093 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a48646ef-6b25-4cef-b573-dae2ff445ce1\") pod \"ovsdbserver-sb-1\" (UID: \"bb75f7eb-e857-4e0c-8900-bd9fea672fb5\") " pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.856097 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.869378 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:10 crc kubenswrapper[4989]: I1006 10:07:10.955219 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.127638 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.223349 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.400668 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.506203 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.593332 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 06 10:07:11 crc kubenswrapper[4989]: W1006 10:07:11.601075 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb75f7eb_e857_4e0c_8900_bd9fea672fb5.slice/crio-b48e81d6e0ce5951c5d40905417bcd61da0a058ee33b5aaa64a4ae945bc8d693 WatchSource:0}: Error finding container b48e81d6e0ce5951c5d40905417bcd61da0a058ee33b5aaa64a4ae945bc8d693: Status 404 returned error can't find the container with id b48e81d6e0ce5951c5d40905417bcd61da0a058ee33b5aaa64a4ae945bc8d693 Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.714955 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"0dabb4ae-66bb-4073-af87-e8a398b22b1a","Type":"ContainerStarted","Data":"07d304591434ff94f3e780e778afaed0d94ec6e4178960dfd23a3de9811b4984"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.715014 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"0dabb4ae-66bb-4073-af87-e8a398b22b1a","Type":"ContainerStarted","Data":"d55e9fa6bfc4f64cee6be65de159e7d9cd16438dcda1558cfdef597b3c66fef6"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.716334 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"27ebef9a-25d5-437a-bd77-5b8ccd0110ba","Type":"ContainerStarted","Data":"51229e6b398f9395743269566595487f9894c4da051759a65698efa0857c6544"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.719675 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"fb3fa8a8-b876-485e-954f-3ed1c711bd61","Type":"ContainerStarted","Data":"9890ee9bb45b803701a102d9f51c2c5d12787be0f2cbff33daf6d3e821468e87"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.728276 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6","Type":"ContainerStarted","Data":"ce2d2b9ee285c99f809b754784695202525157dfcd9e1b9149f0507351a20a96"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.728329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6","Type":"ContainerStarted","Data":"10d66041fc0695ba4ad7a5dc56e27af113314c17d42b85d872864c5c72dbe4d2"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.728345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af55f2bd-22b6-4604-9d04-03dbe5aaf3a6","Type":"ContainerStarted","Data":"7ae3f8377a10ee3a225e403a90acf35a434243c19e72e4036f5879270a5e8e7f"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.731048 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb75f7eb-e857-4e0c-8900-bd9fea672fb5","Type":"ContainerStarted","Data":"b48e81d6e0ce5951c5d40905417bcd61da0a058ee33b5aaa64a4ae945bc8d693"} Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.924758 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.924733724 podStartE2EDuration="2.924733724s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:11.752913145 +0000 UTC m=+5282.542938735" watchObservedRunningTime="2025-10-06 10:07:11.924733724 +0000 UTC m=+5282.714759304" Oct 06 10:07:11 crc kubenswrapper[4989]: I1006 10:07:11.929999 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 06 10:07:11 crc kubenswrapper[4989]: W1006 10:07:11.932946 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc049c234_54cd_4801_9e73_41c5cb8d1404.slice/crio-eb0a5b8f17028f886ebd96b24a4641355c7e167ba5ca1bb73dd438e88e850a8e WatchSource:0}: Error finding container eb0a5b8f17028f886ebd96b24a4641355c7e167ba5ca1bb73dd438e88e850a8e: Status 404 returned error can't find the container with id eb0a5b8f17028f886ebd96b24a4641355c7e167ba5ca1bb73dd438e88e850a8e Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.765839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"0dabb4ae-66bb-4073-af87-e8a398b22b1a","Type":"ContainerStarted","Data":"5624b78eb3e6ef52462a5983aed95d1273c3fc63713561cfa770b595c6b99af3"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.771650 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"27ebef9a-25d5-437a-bd77-5b8ccd0110ba","Type":"ContainerStarted","Data":"5348c0d3efbf01d52765a9136ca18c3ab12676d9337d7ce3ed686249169bc3e2"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.771760 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"27ebef9a-25d5-437a-bd77-5b8ccd0110ba","Type":"ContainerStarted","Data":"79616a30a39b3157c7dddad54f906ae0164539864fa0eeb1b927c84a7cb5c01a"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.775691 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"c049c234-54cd-4801-9e73-41c5cb8d1404","Type":"ContainerStarted","Data":"452aa4026fb3e2f74c3ff82647c94d51af1bea64cf1a1e507c419d585eac217f"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.775799 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"c049c234-54cd-4801-9e73-41c5cb8d1404","Type":"ContainerStarted","Data":"22eea94d4f241424e00568937cd74204b707f4f9fa918b9185457c3a24c0be81"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.775825 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"c049c234-54cd-4801-9e73-41c5cb8d1404","Type":"ContainerStarted","Data":"eb0a5b8f17028f886ebd96b24a4641355c7e167ba5ca1bb73dd438e88e850a8e"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.779564 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"fb3fa8a8-b876-485e-954f-3ed1c711bd61","Type":"ContainerStarted","Data":"33c1fa83e8549b23936c09a581b569e421724d9ec5fd3c906b74bcf3ea745d14"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.779761 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"fb3fa8a8-b876-485e-954f-3ed1c711bd61","Type":"ContainerStarted","Data":"2159643e4ec71634a9dcae8a79536e9021d7c75db5cf0fcb6c5ac7d370e0f5e1"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.782771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb75f7eb-e857-4e0c-8900-bd9fea672fb5","Type":"ContainerStarted","Data":"28dda608464e1c178a95f68e7fca990e237bfe56b42eb4a5f7e8744900297489"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.782811 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"bb75f7eb-e857-4e0c-8900-bd9fea672fb5","Type":"ContainerStarted","Data":"251ed908a8252154b5f4ca03630687d4e5814c4a2cda2484bb1710e349d85a30"} Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.807483 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.807453437 podStartE2EDuration="3.807453437s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:12.803647738 +0000 UTC m=+5283.593673408" watchObservedRunningTime="2025-10-06 10:07:12.807453437 +0000 UTC m=+5283.597479057" Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.828448 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.8284229610000002 podStartE2EDuration="3.828422961s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:12.826065463 +0000 UTC m=+5283.616091073" watchObservedRunningTime="2025-10-06 10:07:12.828422961 +0000 UTC m=+5283.618448561" Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.850789 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.850768585 podStartE2EDuration="3.850768585s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:12.84957049 +0000 UTC m=+5283.639596080" watchObservedRunningTime="2025-10-06 10:07:12.850768585 +0000 UTC m=+5283.640794175" Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.878399 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.87836885 podStartE2EDuration="3.87836885s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:12.874154578 +0000 UTC m=+5283.664180158" watchObservedRunningTime="2025-10-06 10:07:12.87836885 +0000 UTC m=+5283.668394470" Oct 06 10:07:12 crc kubenswrapper[4989]: I1006 10:07:12.902805 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.9027764129999998 podStartE2EDuration="3.902776413s" podCreationTimestamp="2025-10-06 10:07:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:12.892214388 +0000 UTC m=+5283.682239968" watchObservedRunningTime="2025-10-06 10:07:12.902776413 +0000 UTC m=+5283.692802033" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.550037 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.578671 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.593891 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.857685 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.870264 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:13 crc kubenswrapper[4989]: I1006 10:07:13.956043 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.549509 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.578322 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.594110 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.857469 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.870738 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:15 crc kubenswrapper[4989]: I1006 10:07:15.956386 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.625977 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.653602 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.669434 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.675307 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.879986 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.881542 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.903108 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.905968 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.914959 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.920808 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.958919 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:16 crc kubenswrapper[4989]: I1006 10:07:16.968378 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.046455 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.046512 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.046677 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mp74\" (UniqueName: \"kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.046730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.060905 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.148969 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.149014 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.149142 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mp74\" (UniqueName: \"kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.149183 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.152274 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.152301 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.152378 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.185397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mp74\" (UniqueName: \"kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74\") pod \"dnsmasq-dns-dff659fb9-lmwj5\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.210270 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.212744 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.265472 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.384470 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.419590 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.421140 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.424333 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.438310 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.456018 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.456057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j2tv\" (UniqueName: \"kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.456078 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.456106 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.456271 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.557689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.557739 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j2tv\" (UniqueName: \"kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.557761 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.557791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.558782 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.558846 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.558850 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.558888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.559399 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.581429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j2tv\" (UniqueName: \"kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv\") pod \"dnsmasq-dns-7bd564475f-9r8gb\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.755982 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.769332 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:17 crc kubenswrapper[4989]: I1006 10:07:17.880958 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" event={"ID":"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d","Type":"ContainerStarted","Data":"81ae7ae9eae87ec7ee5a3c1970cf874a77cc9ecbdf5b7a1a4ff0888ad21c3131"} Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.265503 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:07:18 crc kubenswrapper[4989]: W1006 10:07:18.273446 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1de44769_2d4c_44eb_9ac9_3e4e90ecfcc5.slice/crio-af368d1327acf616c2071b1894666dd4ee5ad66ad00de4e6132d72ae033a76bc WatchSource:0}: Error finding container af368d1327acf616c2071b1894666dd4ee5ad66ad00de4e6132d72ae033a76bc: Status 404 returned error can't find the container with id af368d1327acf616c2071b1894666dd4ee5ad66ad00de4e6132d72ae033a76bc Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.897704 4989 generic.go:334] "Generic (PLEG): container finished" podID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerID="3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009" exitCode=0 Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.897848 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" event={"ID":"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5","Type":"ContainerDied","Data":"3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009"} Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.898160 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" event={"ID":"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5","Type":"ContainerStarted","Data":"af368d1327acf616c2071b1894666dd4ee5ad66ad00de4e6132d72ae033a76bc"} Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.901869 4989 generic.go:334] "Generic (PLEG): container finished" podID="a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" containerID="7d4f4d85e1e17ac4b4db918162562c539e56aafed79ac6a3c55b9f3d673b6942" exitCode=0 Oct 06 10:07:18 crc kubenswrapper[4989]: I1006 10:07:18.901956 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" event={"ID":"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d","Type":"ContainerDied","Data":"7d4f4d85e1e17ac4b4db918162562c539e56aafed79ac6a3c55b9f3d673b6942"} Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.208125 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.291884 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb\") pod \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.292138 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config\") pod \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.292191 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc\") pod \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.292226 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mp74\" (UniqueName: \"kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74\") pod \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\" (UID: \"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d\") " Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.299275 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74" (OuterVolumeSpecName: "kube-api-access-9mp74") pod "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" (UID: "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d"). InnerVolumeSpecName "kube-api-access-9mp74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.312771 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config" (OuterVolumeSpecName: "config") pod "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" (UID: "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.315971 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" (UID: "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.332183 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" (UID: "a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.394365 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.394681 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.394715 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mp74\" (UniqueName: \"kubernetes.io/projected/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-kube-api-access-9mp74\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.394728 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.934009 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" event={"ID":"a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d","Type":"ContainerDied","Data":"81ae7ae9eae87ec7ee5a3c1970cf874a77cc9ecbdf5b7a1a4ff0888ad21c3131"} Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.934273 4989 scope.go:117] "RemoveContainer" containerID="7d4f4d85e1e17ac4b4db918162562c539e56aafed79ac6a3c55b9f3d673b6942" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.934084 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff659fb9-lmwj5" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.951427 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.951470 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" event={"ID":"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5","Type":"ContainerStarted","Data":"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8"} Oct 06 10:07:19 crc kubenswrapper[4989]: I1006 10:07:19.970824 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" podStartSLOduration=2.97080727 podStartE2EDuration="2.97080727s" podCreationTimestamp="2025-10-06 10:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:19.966688001 +0000 UTC m=+5290.756713591" watchObservedRunningTime="2025-10-06 10:07:19.97080727 +0000 UTC m=+5290.760832860" Oct 06 10:07:20 crc kubenswrapper[4989]: I1006 10:07:20.011941 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:20 crc kubenswrapper[4989]: I1006 10:07:20.023072 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dff659fb9-lmwj5"] Oct 06 10:07:20 crc kubenswrapper[4989]: I1006 10:07:20.633618 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 06 10:07:21 crc kubenswrapper[4989]: I1006 10:07:21.952726 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" path="/var/lib/kubelet/pods/a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d/volumes" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.866246 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 06 10:07:23 crc kubenswrapper[4989]: E1006 10:07:23.867209 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" containerName="init" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.867233 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" containerName="init" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.867539 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a80a7c6c-e75c-4c3d-8f9b-2aee7ac6799d" containerName="init" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.868564 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.870822 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.885193 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.981766 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.981887 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:23 crc kubenswrapper[4989]: I1006 10:07:23.981959 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ltxj\" (UniqueName: \"kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.085295 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.086255 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.086296 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ltxj\" (UniqueName: \"kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.090284 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.090331 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c58b5be0376ac9cf17973e8fb986ffe69fc12916d29dbaf6922160b0326d795c/globalmount\"" pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.100326 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.108365 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ltxj\" (UniqueName: \"kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.143612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") pod \"ovn-copy-data\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.205237 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.623917 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.996615 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"30510729-0329-47b2-ba21-8133a7559f05","Type":"ContainerStarted","Data":"cd30673b163ccb8c4145a865b292e380edf28f5a7a31924e7dff26929382a556"} Oct 06 10:07:24 crc kubenswrapper[4989]: I1006 10:07:24.997048 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"30510729-0329-47b2-ba21-8133a7559f05","Type":"ContainerStarted","Data":"62311dd598feeac8d4ff0fef8990a84cbe3ba8efc84d5e36d4a51a8c05de7cac"} Oct 06 10:07:25 crc kubenswrapper[4989]: I1006 10:07:25.014411 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.01439129 podStartE2EDuration="3.01439129s" podCreationTimestamp="2025-10-06 10:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:25.012066063 +0000 UTC m=+5295.802091653" watchObservedRunningTime="2025-10-06 10:07:25.01439129 +0000 UTC m=+5295.804416870" Oct 06 10:07:27 crc kubenswrapper[4989]: I1006 10:07:27.758908 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:07:27 crc kubenswrapper[4989]: I1006 10:07:27.839027 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:07:27 crc kubenswrapper[4989]: I1006 10:07:27.839467 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="dnsmasq-dns" containerID="cri-o://5ad9d0aaa98c0398181a67bd203750809ea27633ee08b86391fd2084efbf8bed" gracePeriod=10 Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.031118 4989 generic.go:334] "Generic (PLEG): container finished" podID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerID="5ad9d0aaa98c0398181a67bd203750809ea27633ee08b86391fd2084efbf8bed" exitCode=0 Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.031159 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" event={"ID":"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0","Type":"ContainerDied","Data":"5ad9d0aaa98c0398181a67bd203750809ea27633ee08b86391fd2084efbf8bed"} Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.295470 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.372820 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc\") pod \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.372938 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwrvq\" (UniqueName: \"kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq\") pod \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.373063 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config\") pod \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\" (UID: \"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0\") " Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.378922 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq" (OuterVolumeSpecName: "kube-api-access-kwrvq") pod "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" (UID: "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0"). InnerVolumeSpecName "kube-api-access-kwrvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.415350 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" (UID: "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.425299 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config" (OuterVolumeSpecName: "config") pod "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" (UID: "e9c8b79e-f897-49a3-8ce3-4d7630b55ae0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.475325 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.475361 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:28 crc kubenswrapper[4989]: I1006 10:07:28.475373 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwrvq\" (UniqueName: \"kubernetes.io/projected/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0-kube-api-access-kwrvq\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.042014 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" event={"ID":"e9c8b79e-f897-49a3-8ce3-4d7630b55ae0","Type":"ContainerDied","Data":"269f67ca234b82c77037ed594d9b258b2b6fa4e092e4248ede38378d86eb639e"} Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.042075 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-f2bjf" Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.042100 4989 scope.go:117] "RemoveContainer" containerID="5ad9d0aaa98c0398181a67bd203750809ea27633ee08b86391fd2084efbf8bed" Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.064591 4989 scope.go:117] "RemoveContainer" containerID="96862c2ec866b85a8607626cf1bc9890315efcd53d51ea77318825a1c03b2840" Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.080215 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.088083 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-f2bjf"] Oct 06 10:07:29 crc kubenswrapper[4989]: I1006 10:07:29.951482 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" path="/var/lib/kubelet/pods/e9c8b79e-f897-49a3-8ce3-4d7630b55ae0/volumes" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.082073 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:07:31 crc kubenswrapper[4989]: E1006 10:07:31.082805 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="dnsmasq-dns" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.082826 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="dnsmasq-dns" Oct 06 10:07:31 crc kubenswrapper[4989]: E1006 10:07:31.082858 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="init" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.082869 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="init" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.083117 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c8b79e-f897-49a3-8ce3-4d7630b55ae0" containerName="dnsmasq-dns" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.084877 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.089350 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.100373 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.101891 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.102030 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5g2zw" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.225487 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-config\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.225842 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e62e912-c83a-4f76-904e-45d246480cc6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.225877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-scripts\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.225925 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e62e912-c83a-4f76-904e-45d246480cc6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.225946 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klfnd\" (UniqueName: \"kubernetes.io/projected/6e62e912-c83a-4f76-904e-45d246480cc6-kube-api-access-klfnd\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327030 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e62e912-c83a-4f76-904e-45d246480cc6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klfnd\" (UniqueName: \"kubernetes.io/projected/6e62e912-c83a-4f76-904e-45d246480cc6-kube-api-access-klfnd\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327164 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-config\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327181 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e62e912-c83a-4f76-904e-45d246480cc6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327209 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-scripts\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.327779 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e62e912-c83a-4f76-904e-45d246480cc6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.328141 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-scripts\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.328159 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e62e912-c83a-4f76-904e-45d246480cc6-config\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.334168 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e62e912-c83a-4f76-904e-45d246480cc6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.343871 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klfnd\" (UniqueName: \"kubernetes.io/projected/6e62e912-c83a-4f76-904e-45d246480cc6-kube-api-access-klfnd\") pod \"ovn-northd-0\" (UID: \"6e62e912-c83a-4f76-904e-45d246480cc6\") " pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.442454 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 06 10:07:31 crc kubenswrapper[4989]: I1006 10:07:31.886573 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 06 10:07:31 crc kubenswrapper[4989]: W1006 10:07:31.895842 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e62e912_c83a_4f76_904e_45d246480cc6.slice/crio-d4a5ea55049ea01b40906a3a82ddaf2e5469732e0fc9c7104e429d4afcc2f040 WatchSource:0}: Error finding container d4a5ea55049ea01b40906a3a82ddaf2e5469732e0fc9c7104e429d4afcc2f040: Status 404 returned error can't find the container with id d4a5ea55049ea01b40906a3a82ddaf2e5469732e0fc9c7104e429d4afcc2f040 Oct 06 10:07:32 crc kubenswrapper[4989]: I1006 10:07:32.089681 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6e62e912-c83a-4f76-904e-45d246480cc6","Type":"ContainerStarted","Data":"8760768cc90ad2aa35d622a0a227853d927decf05f51a23ceb13e65d30ff02c3"} Oct 06 10:07:32 crc kubenswrapper[4989]: I1006 10:07:32.090020 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6e62e912-c83a-4f76-904e-45d246480cc6","Type":"ContainerStarted","Data":"d4a5ea55049ea01b40906a3a82ddaf2e5469732e0fc9c7104e429d4afcc2f040"} Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.106550 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"6e62e912-c83a-4f76-904e-45d246480cc6","Type":"ContainerStarted","Data":"6ccd0bd8b65bd66c404440c358a7adcaca8e1255e4f76242c04e284d02b66b94"} Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.107015 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.133545 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.133514729 podStartE2EDuration="2.133514729s" podCreationTimestamp="2025-10-06 10:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:33.129940346 +0000 UTC m=+5303.919965986" watchObservedRunningTime="2025-10-06 10:07:33.133514729 +0000 UTC m=+5303.923540349" Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.935979 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.936341 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.955459 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.956791 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:07:33 crc kubenswrapper[4989]: I1006 10:07:33.956953 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" gracePeriod=600 Oct 06 10:07:34 crc kubenswrapper[4989]: E1006 10:07:34.098137 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:07:34 crc kubenswrapper[4989]: I1006 10:07:34.124695 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" exitCode=0 Oct 06 10:07:34 crc kubenswrapper[4989]: I1006 10:07:34.124742 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5"} Oct 06 10:07:34 crc kubenswrapper[4989]: I1006 10:07:34.124819 4989 scope.go:117] "RemoveContainer" containerID="c442e72f9b31f7ade4584c05a9c18b838a8a5e6a4a7bcd3957950c7e09aca69b" Oct 06 10:07:34 crc kubenswrapper[4989]: I1006 10:07:34.125904 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:07:34 crc kubenswrapper[4989]: E1006 10:07:34.126280 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.545906 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ctp2p"] Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.547682 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.556896 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ctp2p"] Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.632726 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqz8q\" (UniqueName: \"kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q\") pod \"keystone-db-create-ctp2p\" (UID: \"26f40369-b4d3-46c1-84e2-21d7e34b0d7f\") " pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.733926 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqz8q\" (UniqueName: \"kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q\") pod \"keystone-db-create-ctp2p\" (UID: \"26f40369-b4d3-46c1-84e2-21d7e34b0d7f\") " pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.752509 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqz8q\" (UniqueName: \"kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q\") pod \"keystone-db-create-ctp2p\" (UID: \"26f40369-b4d3-46c1-84e2-21d7e34b0d7f\") " pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:36 crc kubenswrapper[4989]: I1006 10:07:36.870245 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:37 crc kubenswrapper[4989]: I1006 10:07:37.395355 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ctp2p"] Oct 06 10:07:37 crc kubenswrapper[4989]: W1006 10:07:37.407597 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26f40369_b4d3_46c1_84e2_21d7e34b0d7f.slice/crio-d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06 WatchSource:0}: Error finding container d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06: Status 404 returned error can't find the container with id d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06 Oct 06 10:07:38 crc kubenswrapper[4989]: I1006 10:07:38.167826 4989 generic.go:334] "Generic (PLEG): container finished" podID="26f40369-b4d3-46c1-84e2-21d7e34b0d7f" containerID="82e3fb0a8e9d221cf3a76861dda7db006629cd32643acd24cc14bde43d65fd41" exitCode=0 Oct 06 10:07:38 crc kubenswrapper[4989]: I1006 10:07:38.167896 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ctp2p" event={"ID":"26f40369-b4d3-46c1-84e2-21d7e34b0d7f","Type":"ContainerDied","Data":"82e3fb0a8e9d221cf3a76861dda7db006629cd32643acd24cc14bde43d65fd41"} Oct 06 10:07:38 crc kubenswrapper[4989]: I1006 10:07:38.167938 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ctp2p" event={"ID":"26f40369-b4d3-46c1-84e2-21d7e34b0d7f","Type":"ContainerStarted","Data":"d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06"} Oct 06 10:07:39 crc kubenswrapper[4989]: I1006 10:07:39.579149 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:39 crc kubenswrapper[4989]: I1006 10:07:39.687125 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqz8q\" (UniqueName: \"kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q\") pod \"26f40369-b4d3-46c1-84e2-21d7e34b0d7f\" (UID: \"26f40369-b4d3-46c1-84e2-21d7e34b0d7f\") " Oct 06 10:07:39 crc kubenswrapper[4989]: I1006 10:07:39.695892 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q" (OuterVolumeSpecName: "kube-api-access-gqz8q") pod "26f40369-b4d3-46c1-84e2-21d7e34b0d7f" (UID: "26f40369-b4d3-46c1-84e2-21d7e34b0d7f"). InnerVolumeSpecName "kube-api-access-gqz8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:39 crc kubenswrapper[4989]: I1006 10:07:39.789137 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqz8q\" (UniqueName: \"kubernetes.io/projected/26f40369-b4d3-46c1-84e2-21d7e34b0d7f-kube-api-access-gqz8q\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:40 crc kubenswrapper[4989]: I1006 10:07:40.196153 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ctp2p" event={"ID":"26f40369-b4d3-46c1-84e2-21d7e34b0d7f","Type":"ContainerDied","Data":"d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06"} Oct 06 10:07:40 crc kubenswrapper[4989]: I1006 10:07:40.196230 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d14ca65ac7a8f794c0982fdbe2bca7ff1ed3e43e3e380a03a9abccbb8ee65e06" Oct 06 10:07:40 crc kubenswrapper[4989]: I1006 10:07:40.196335 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ctp2p" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.507600 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.658240 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8171-account-create-xvtbw"] Oct 06 10:07:46 crc kubenswrapper[4989]: E1006 10:07:46.658619 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f40369-b4d3-46c1-84e2-21d7e34b0d7f" containerName="mariadb-database-create" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.658632 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f40369-b4d3-46c1-84e2-21d7e34b0d7f" containerName="mariadb-database-create" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.658837 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f40369-b4d3-46c1-84e2-21d7e34b0d7f" containerName="mariadb-database-create" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.659363 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.661426 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.666927 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8171-account-create-xvtbw"] Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.720941 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b56nx\" (UniqueName: \"kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx\") pod \"keystone-8171-account-create-xvtbw\" (UID: \"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5\") " pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.822368 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b56nx\" (UniqueName: \"kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx\") pod \"keystone-8171-account-create-xvtbw\" (UID: \"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5\") " pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.844183 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b56nx\" (UniqueName: \"kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx\") pod \"keystone-8171-account-create-xvtbw\" (UID: \"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5\") " pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.936163 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:07:46 crc kubenswrapper[4989]: E1006 10:07:46.936739 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:07:46 crc kubenswrapper[4989]: I1006 10:07:46.978014 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:47 crc kubenswrapper[4989]: I1006 10:07:47.464360 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8171-account-create-xvtbw"] Oct 06 10:07:47 crc kubenswrapper[4989]: W1006 10:07:47.471810 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb724a6_b54c_42ae_bcf5_a6fae28d19c5.slice/crio-d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08 WatchSource:0}: Error finding container d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08: Status 404 returned error can't find the container with id d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08 Oct 06 10:07:48 crc kubenswrapper[4989]: I1006 10:07:48.288755 4989 generic.go:334] "Generic (PLEG): container finished" podID="1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" containerID="169fe97811ea870cbd51b9bdbc1d3ed8cf6d74b9fe1a49516f4396dddf7cb082" exitCode=0 Oct 06 10:07:48 crc kubenswrapper[4989]: I1006 10:07:48.288865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8171-account-create-xvtbw" event={"ID":"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5","Type":"ContainerDied","Data":"169fe97811ea870cbd51b9bdbc1d3ed8cf6d74b9fe1a49516f4396dddf7cb082"} Oct 06 10:07:48 crc kubenswrapper[4989]: I1006 10:07:48.289235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8171-account-create-xvtbw" event={"ID":"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5","Type":"ContainerStarted","Data":"d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08"} Oct 06 10:07:49 crc kubenswrapper[4989]: I1006 10:07:49.653173 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:49 crc kubenswrapper[4989]: I1006 10:07:49.796788 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b56nx\" (UniqueName: \"kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx\") pod \"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5\" (UID: \"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5\") " Oct 06 10:07:49 crc kubenswrapper[4989]: I1006 10:07:49.805886 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx" (OuterVolumeSpecName: "kube-api-access-b56nx") pod "1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" (UID: "1eb724a6-b54c-42ae-bcf5-a6fae28d19c5"). InnerVolumeSpecName "kube-api-access-b56nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:49 crc kubenswrapper[4989]: I1006 10:07:49.899345 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b56nx\" (UniqueName: \"kubernetes.io/projected/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5-kube-api-access-b56nx\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:50 crc kubenswrapper[4989]: I1006 10:07:50.312474 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8171-account-create-xvtbw" event={"ID":"1eb724a6-b54c-42ae-bcf5-a6fae28d19c5","Type":"ContainerDied","Data":"d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08"} Oct 06 10:07:50 crc kubenswrapper[4989]: I1006 10:07:50.312534 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7a05bb5c86ad019f9a00f81c65adec6a6d75ce09419d7c3ad285bcf2abafa08" Oct 06 10:07:50 crc kubenswrapper[4989]: I1006 10:07:50.312576 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8171-account-create-xvtbw" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.049746 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jfp5d"] Oct 06 10:07:52 crc kubenswrapper[4989]: E1006 10:07:52.051202 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" containerName="mariadb-account-create" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.051271 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" containerName="mariadb-account-create" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.051509 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" containerName="mariadb-account-create" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.052078 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.054153 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.054676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.054785 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4z52d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.054843 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.060666 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jfp5d"] Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.140293 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.140453 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bgr4\" (UniqueName: \"kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.140608 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.241707 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bgr4\" (UniqueName: \"kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.241812 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.241868 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.249681 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.256208 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.258408 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bgr4\" (UniqueName: \"kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4\") pod \"keystone-db-sync-jfp5d\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.370623 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:52 crc kubenswrapper[4989]: I1006 10:07:52.943948 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jfp5d"] Oct 06 10:07:53 crc kubenswrapper[4989]: I1006 10:07:53.347358 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jfp5d" event={"ID":"91f732e7-a08f-41b9-80fc-bcf60cb142d8","Type":"ContainerStarted","Data":"9a527d4b848b349bde90deaa00816cde1b68e159e0d5f9f914a7cf564bbd9584"} Oct 06 10:07:53 crc kubenswrapper[4989]: I1006 10:07:53.347838 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jfp5d" event={"ID":"91f732e7-a08f-41b9-80fc-bcf60cb142d8","Type":"ContainerStarted","Data":"1aa3d7f46acee231d39a3a7c7bb5a5c34066c24314715897ec0c423fda2e0c72"} Oct 06 10:07:53 crc kubenswrapper[4989]: I1006 10:07:53.373684 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jfp5d" podStartSLOduration=1.373632896 podStartE2EDuration="1.373632896s" podCreationTimestamp="2025-10-06 10:07:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:53.366907992 +0000 UTC m=+5324.156933622" watchObservedRunningTime="2025-10-06 10:07:53.373632896 +0000 UTC m=+5324.163658506" Oct 06 10:07:55 crc kubenswrapper[4989]: I1006 10:07:55.376293 4989 generic.go:334] "Generic (PLEG): container finished" podID="91f732e7-a08f-41b9-80fc-bcf60cb142d8" containerID="9a527d4b848b349bde90deaa00816cde1b68e159e0d5f9f914a7cf564bbd9584" exitCode=0 Oct 06 10:07:55 crc kubenswrapper[4989]: I1006 10:07:55.376584 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jfp5d" event={"ID":"91f732e7-a08f-41b9-80fc-bcf60cb142d8","Type":"ContainerDied","Data":"9a527d4b848b349bde90deaa00816cde1b68e159e0d5f9f914a7cf564bbd9584"} Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.778809 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.937374 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data\") pod \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.937481 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bgr4\" (UniqueName: \"kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4\") pod \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.937590 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle\") pod \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\" (UID: \"91f732e7-a08f-41b9-80fc-bcf60cb142d8\") " Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.945352 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4" (OuterVolumeSpecName: "kube-api-access-8bgr4") pod "91f732e7-a08f-41b9-80fc-bcf60cb142d8" (UID: "91f732e7-a08f-41b9-80fc-bcf60cb142d8"). InnerVolumeSpecName "kube-api-access-8bgr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:07:56 crc kubenswrapper[4989]: I1006 10:07:56.968186 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91f732e7-a08f-41b9-80fc-bcf60cb142d8" (UID: "91f732e7-a08f-41b9-80fc-bcf60cb142d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.018760 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data" (OuterVolumeSpecName: "config-data") pod "91f732e7-a08f-41b9-80fc-bcf60cb142d8" (UID: "91f732e7-a08f-41b9-80fc-bcf60cb142d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.040634 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.040715 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f732e7-a08f-41b9-80fc-bcf60cb142d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.040737 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bgr4\" (UniqueName: \"kubernetes.io/projected/91f732e7-a08f-41b9-80fc-bcf60cb142d8-kube-api-access-8bgr4\") on node \"crc\" DevicePath \"\"" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.399479 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jfp5d" event={"ID":"91f732e7-a08f-41b9-80fc-bcf60cb142d8","Type":"ContainerDied","Data":"1aa3d7f46acee231d39a3a7c7bb5a5c34066c24314715897ec0c423fda2e0c72"} Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.399542 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aa3d7f46acee231d39a3a7c7bb5a5c34066c24314715897ec0c423fda2e0c72" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.399563 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jfp5d" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.684303 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:07:57 crc kubenswrapper[4989]: E1006 10:07:57.684612 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f732e7-a08f-41b9-80fc-bcf60cb142d8" containerName="keystone-db-sync" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.684628 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f732e7-a08f-41b9-80fc-bcf60cb142d8" containerName="keystone-db-sync" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.684837 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f732e7-a08f-41b9-80fc-bcf60cb142d8" containerName="keystone-db-sync" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.685602 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.699491 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bjbmt"] Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.700845 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.702928 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.703078 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.703180 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.703482 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4z52d" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.732492 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.748428 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bjbmt"] Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855412 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855794 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855830 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855850 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.855899 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.856112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgg75\" (UniqueName: \"kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.856275 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.856305 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.856372 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjtzl\" (UniqueName: \"kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.856402 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.957926 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958052 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958095 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958119 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958144 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958169 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958205 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgg75\" (UniqueName: \"kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958251 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958273 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958315 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjtzl\" (UniqueName: \"kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.958340 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.960446 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.960569 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.960886 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.961243 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.966328 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.966541 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.967039 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.968209 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.978135 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.979497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjtzl\" (UniqueName: \"kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl\") pod \"keystone-bootstrap-bjbmt\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:57 crc kubenswrapper[4989]: I1006 10:07:57.980932 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgg75\" (UniqueName: \"kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75\") pod \"dnsmasq-dns-85d4b57f6f-j7r5m\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:58 crc kubenswrapper[4989]: I1006 10:07:58.003209 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:07:58 crc kubenswrapper[4989]: I1006 10:07:58.024710 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:07:58 crc kubenswrapper[4989]: I1006 10:07:58.482493 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:07:58 crc kubenswrapper[4989]: I1006 10:07:58.559006 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bjbmt"] Oct 06 10:07:58 crc kubenswrapper[4989]: I1006 10:07:58.937360 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:07:58 crc kubenswrapper[4989]: E1006 10:07:58.938111 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.418219 4989 generic.go:334] "Generic (PLEG): container finished" podID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerID="c1e5b7c39d73b84bac99e546e3a04747c71ba8feca977ca16e28ab9ef7ac316e" exitCode=0 Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.418280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" event={"ID":"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a","Type":"ContainerDied","Data":"c1e5b7c39d73b84bac99e546e3a04747c71ba8feca977ca16e28ab9ef7ac316e"} Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.418304 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" event={"ID":"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a","Type":"ContainerStarted","Data":"26a0ce4cf87509c345da12e1b6cf44ceaca1b67decd05ad310c8f1454e628ca3"} Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.421336 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bjbmt" event={"ID":"fff4589b-f50c-48d7-97ed-626929e222dd","Type":"ContainerStarted","Data":"067666b00022b9c722ada3f7ecf2a1b88212fd231758196c1d85dbce67768234"} Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.421411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bjbmt" event={"ID":"fff4589b-f50c-48d7-97ed-626929e222dd","Type":"ContainerStarted","Data":"04823269d5e773ba6fa1ac5c95bd7df5bc70c731efc826fe8d2e438f848e8c84"} Oct 06 10:07:59 crc kubenswrapper[4989]: I1006 10:07:59.483834 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bjbmt" podStartSLOduration=2.483805995 podStartE2EDuration="2.483805995s" podCreationTimestamp="2025-10-06 10:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:07:59.472907951 +0000 UTC m=+5330.262933571" watchObservedRunningTime="2025-10-06 10:07:59.483805995 +0000 UTC m=+5330.273831615" Oct 06 10:08:00 crc kubenswrapper[4989]: I1006 10:08:00.438054 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" event={"ID":"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a","Type":"ContainerStarted","Data":"891fa4f95e00f2312f6f291e60e13aeb9696640b2d2587c1a29020f62707e947"} Oct 06 10:08:00 crc kubenswrapper[4989]: I1006 10:08:00.439595 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:08:00 crc kubenswrapper[4989]: I1006 10:08:00.472544 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" podStartSLOduration=3.472527702 podStartE2EDuration="3.472527702s" podCreationTimestamp="2025-10-06 10:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:08:00.467411115 +0000 UTC m=+5331.257436695" watchObservedRunningTime="2025-10-06 10:08:00.472527702 +0000 UTC m=+5331.262553282" Oct 06 10:08:03 crc kubenswrapper[4989]: I1006 10:08:03.478847 4989 generic.go:334] "Generic (PLEG): container finished" podID="fff4589b-f50c-48d7-97ed-626929e222dd" containerID="067666b00022b9c722ada3f7ecf2a1b88212fd231758196c1d85dbce67768234" exitCode=0 Oct 06 10:08:03 crc kubenswrapper[4989]: I1006 10:08:03.479013 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bjbmt" event={"ID":"fff4589b-f50c-48d7-97ed-626929e222dd","Type":"ContainerDied","Data":"067666b00022b9c722ada3f7ecf2a1b88212fd231758196c1d85dbce67768234"} Oct 06 10:08:04 crc kubenswrapper[4989]: I1006 10:08:04.852399 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.004953 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjtzl\" (UniqueName: \"kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.005030 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.005079 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.005120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.005185 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.005225 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle\") pod \"fff4589b-f50c-48d7-97ed-626929e222dd\" (UID: \"fff4589b-f50c-48d7-97ed-626929e222dd\") " Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.011829 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.012568 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.012686 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts" (OuterVolumeSpecName: "scripts") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.014058 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl" (OuterVolumeSpecName: "kube-api-access-jjtzl") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "kube-api-access-jjtzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.041588 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.054884 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data" (OuterVolumeSpecName: "config-data") pod "fff4589b-f50c-48d7-97ed-626929e222dd" (UID: "fff4589b-f50c-48d7-97ed-626929e222dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108054 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108310 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108501 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108605 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108723 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fff4589b-f50c-48d7-97ed-626929e222dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.108839 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjtzl\" (UniqueName: \"kubernetes.io/projected/fff4589b-f50c-48d7-97ed-626929e222dd-kube-api-access-jjtzl\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.500716 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bjbmt" event={"ID":"fff4589b-f50c-48d7-97ed-626929e222dd","Type":"ContainerDied","Data":"04823269d5e773ba6fa1ac5c95bd7df5bc70c731efc826fe8d2e438f848e8c84"} Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.500806 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04823269d5e773ba6fa1ac5c95bd7df5bc70c731efc826fe8d2e438f848e8c84" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.500805 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bjbmt" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.603713 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bjbmt"] Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.614796 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bjbmt"] Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.679410 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rlv82"] Oct 06 10:08:05 crc kubenswrapper[4989]: E1006 10:08:05.679918 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fff4589b-f50c-48d7-97ed-626929e222dd" containerName="keystone-bootstrap" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.679941 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fff4589b-f50c-48d7-97ed-626929e222dd" containerName="keystone-bootstrap" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.680139 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fff4589b-f50c-48d7-97ed-626929e222dd" containerName="keystone-bootstrap" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.680867 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.683030 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.684676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4z52d" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.685639 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.685825 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.687741 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rlv82"] Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.820709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.820828 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.820931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.820957 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbx9g\" (UniqueName: \"kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.820982 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.821013 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922281 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922357 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922375 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbx9g\" (UniqueName: \"kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922393 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922411 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.922436 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.927161 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.927758 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.928127 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.928530 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.929671 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.942216 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbx9g\" (UniqueName: \"kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g\") pod \"keystone-bootstrap-rlv82\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:05 crc kubenswrapper[4989]: I1006 10:08:05.952086 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fff4589b-f50c-48d7-97ed-626929e222dd" path="/var/lib/kubelet/pods/fff4589b-f50c-48d7-97ed-626929e222dd/volumes" Oct 06 10:08:06 crc kubenswrapper[4989]: I1006 10:08:06.023493 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:06 crc kubenswrapper[4989]: I1006 10:08:06.466773 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rlv82"] Oct 06 10:08:06 crc kubenswrapper[4989]: I1006 10:08:06.519202 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rlv82" event={"ID":"adee3030-b6d6-4ba9-9284-c5b62de44d46","Type":"ContainerStarted","Data":"4a1463a2f1e5fcde72b5dea43b95286f8232cb085ea074d6640ae407c734e199"} Oct 06 10:08:07 crc kubenswrapper[4989]: I1006 10:08:07.539183 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rlv82" event={"ID":"adee3030-b6d6-4ba9-9284-c5b62de44d46","Type":"ContainerStarted","Data":"e4ced04d64dc10f837231c9958a4e615f3bdee5a1441218894f0339ead0fde15"} Oct 06 10:08:07 crc kubenswrapper[4989]: I1006 10:08:07.571257 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rlv82" podStartSLOduration=2.571233022 podStartE2EDuration="2.571233022s" podCreationTimestamp="2025-10-06 10:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:08:07.56662117 +0000 UTC m=+5338.356646800" watchObservedRunningTime="2025-10-06 10:08:07.571233022 +0000 UTC m=+5338.361258622" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.005830 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.073524 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.073774 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="dnsmasq-dns" containerID="cri-o://1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8" gracePeriod=10 Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.543103 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.554284 4989 generic.go:334] "Generic (PLEG): container finished" podID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerID="1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8" exitCode=0 Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.554356 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.554371 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" event={"ID":"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5","Type":"ContainerDied","Data":"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8"} Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.554480 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd564475f-9r8gb" event={"ID":"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5","Type":"ContainerDied","Data":"af368d1327acf616c2071b1894666dd4ee5ad66ad00de4e6132d72ae033a76bc"} Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.554502 4989 scope.go:117] "RemoveContainer" containerID="1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.575873 4989 scope.go:117] "RemoveContainer" containerID="3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.596843 4989 scope.go:117] "RemoveContainer" containerID="1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8" Oct 06 10:08:08 crc kubenswrapper[4989]: E1006 10:08:08.597345 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8\": container with ID starting with 1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8 not found: ID does not exist" containerID="1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.597373 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8"} err="failed to get container status \"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8\": rpc error: code = NotFound desc = could not find container \"1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8\": container with ID starting with 1dfde93cfcc9a3481a2357ebddce448591652cc814b22f7b934fee5f97dd33c8 not found: ID does not exist" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.597393 4989 scope.go:117] "RemoveContainer" containerID="3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009" Oct 06 10:08:08 crc kubenswrapper[4989]: E1006 10:08:08.597853 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009\": container with ID starting with 3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009 not found: ID does not exist" containerID="3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.597913 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009"} err="failed to get container status \"3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009\": rpc error: code = NotFound desc = could not find container \"3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009\": container with ID starting with 3d39cc0e2bce3f78b825d8c7e6c2e07f6d93099bd310df1ad1136c0efa416009 not found: ID does not exist" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.684850 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config\") pod \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.684937 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb\") pod \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.685844 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc\") pod \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.686024 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j2tv\" (UniqueName: \"kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv\") pod \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.686092 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb\") pod \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\" (UID: \"1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5\") " Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.696892 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv" (OuterVolumeSpecName: "kube-api-access-6j2tv") pod "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" (UID: "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5"). InnerVolumeSpecName "kube-api-access-6j2tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.724630 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config" (OuterVolumeSpecName: "config") pod "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" (UID: "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.725416 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" (UID: "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.726499 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" (UID: "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.734341 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" (UID: "1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.788719 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j2tv\" (UniqueName: \"kubernetes.io/projected/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-kube-api-access-6j2tv\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.788768 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.788788 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.788809 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.788825 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.894352 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:08:08 crc kubenswrapper[4989]: I1006 10:08:08.900012 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd564475f-9r8gb"] Oct 06 10:08:09 crc kubenswrapper[4989]: I1006 10:08:09.568437 4989 generic.go:334] "Generic (PLEG): container finished" podID="adee3030-b6d6-4ba9-9284-c5b62de44d46" containerID="e4ced04d64dc10f837231c9958a4e615f3bdee5a1441218894f0339ead0fde15" exitCode=0 Oct 06 10:08:09 crc kubenswrapper[4989]: I1006 10:08:09.568528 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rlv82" event={"ID":"adee3030-b6d6-4ba9-9284-c5b62de44d46","Type":"ContainerDied","Data":"e4ced04d64dc10f837231c9958a4e615f3bdee5a1441218894f0339ead0fde15"} Oct 06 10:08:09 crc kubenswrapper[4989]: I1006 10:08:09.961575 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" path="/var/lib/kubelet/pods/1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5/volumes" Oct 06 10:08:10 crc kubenswrapper[4989]: I1006 10:08:10.927795 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033381 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033460 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbx9g\" (UniqueName: \"kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033572 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033640 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033675 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.033746 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts\") pod \"adee3030-b6d6-4ba9-9284-c5b62de44d46\" (UID: \"adee3030-b6d6-4ba9-9284-c5b62de44d46\") " Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.039958 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.040463 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g" (OuterVolumeSpecName: "kube-api-access-qbx9g") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "kube-api-access-qbx9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.040828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts" (OuterVolumeSpecName: "scripts") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.041677 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.057687 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.068864 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data" (OuterVolumeSpecName: "config-data") pod "adee3030-b6d6-4ba9-9284-c5b62de44d46" (UID: "adee3030-b6d6-4ba9-9284-c5b62de44d46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135897 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135926 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135938 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135948 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135956 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbx9g\" (UniqueName: \"kubernetes.io/projected/adee3030-b6d6-4ba9-9284-c5b62de44d46-kube-api-access-qbx9g\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.135964 4989 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/adee3030-b6d6-4ba9-9284-c5b62de44d46-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.599759 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rlv82" event={"ID":"adee3030-b6d6-4ba9-9284-c5b62de44d46","Type":"ContainerDied","Data":"4a1463a2f1e5fcde72b5dea43b95286f8232cb085ea074d6640ae407c734e199"} Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.599824 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a1463a2f1e5fcde72b5dea43b95286f8232cb085ea074d6640ae407c734e199" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.599878 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rlv82" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.801712 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-555c6df57c-dz9gw"] Oct 06 10:08:11 crc kubenswrapper[4989]: E1006 10:08:11.802458 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="dnsmasq-dns" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.802597 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="dnsmasq-dns" Oct 06 10:08:11 crc kubenswrapper[4989]: E1006 10:08:11.802843 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="init" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.802990 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="init" Oct 06 10:08:11 crc kubenswrapper[4989]: E1006 10:08:11.803117 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adee3030-b6d6-4ba9-9284-c5b62de44d46" containerName="keystone-bootstrap" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.803240 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="adee3030-b6d6-4ba9-9284-c5b62de44d46" containerName="keystone-bootstrap" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.804232 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de44769-2d4c-44eb-9ac9-3e4e90ecfcc5" containerName="dnsmasq-dns" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.804355 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="adee3030-b6d6-4ba9-9284-c5b62de44d46" containerName="keystone-bootstrap" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.805062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.807199 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.807526 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4z52d" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.807762 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.808713 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.813262 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-555c6df57c-dz9gw"] Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.948676 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-config-data\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.948747 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9ptq\" (UniqueName: \"kubernetes.io/projected/32295746-8fcc-4a25-87e3-75ab110afd5a-kube-api-access-n9ptq\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.948781 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-fernet-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.948919 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-scripts\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.949062 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-credential-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:11 crc kubenswrapper[4989]: I1006 10:08:11.949121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-combined-ca-bundle\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.050719 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-scripts\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.050872 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-credential-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.050948 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-combined-ca-bundle\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.051069 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-config-data\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.051190 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9ptq\" (UniqueName: \"kubernetes.io/projected/32295746-8fcc-4a25-87e3-75ab110afd5a-kube-api-access-n9ptq\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.051282 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-fernet-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.056729 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-combined-ca-bundle\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.057019 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-scripts\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.059197 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-config-data\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.060905 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-fernet-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.063145 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/32295746-8fcc-4a25-87e3-75ab110afd5a-credential-keys\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.085226 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9ptq\" (UniqueName: \"kubernetes.io/projected/32295746-8fcc-4a25-87e3-75ab110afd5a-kube-api-access-n9ptq\") pod \"keystone-555c6df57c-dz9gw\" (UID: \"32295746-8fcc-4a25-87e3-75ab110afd5a\") " pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.127056 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.601303 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-555c6df57c-dz9gw"] Oct 06 10:08:12 crc kubenswrapper[4989]: I1006 10:08:12.936329 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:08:12 crc kubenswrapper[4989]: E1006 10:08:12.937212 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:08:13 crc kubenswrapper[4989]: I1006 10:08:13.627183 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-555c6df57c-dz9gw" event={"ID":"32295746-8fcc-4a25-87e3-75ab110afd5a","Type":"ContainerStarted","Data":"24fbfa6bdd541605d07515a3afdb8dd83ff4f990358431e9d262cfc49e272268"} Oct 06 10:08:13 crc kubenswrapper[4989]: I1006 10:08:13.627667 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:13 crc kubenswrapper[4989]: I1006 10:08:13.627681 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-555c6df57c-dz9gw" event={"ID":"32295746-8fcc-4a25-87e3-75ab110afd5a","Type":"ContainerStarted","Data":"4ccfe086ca854fcc88f0ef5456486c54c7bf7c35d254cb0ea26dc1bec57f96c3"} Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.264177 4989 scope.go:117] "RemoveContainer" containerID="997848024c9b9185ee273592d18e6d5df21125353e0ee828e3bf43b753b11808" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.300454 4989 scope.go:117] "RemoveContainer" containerID="87f7a737a3b525eecca89b0c2ac0dbfce46231f8e326b3f9960964556de8cb03" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.338941 4989 scope.go:117] "RemoveContainer" containerID="c1ac576c81166aecc1caafd707d246cd1cf381748ba8b714806079c6887e020b" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.364103 4989 scope.go:117] "RemoveContainer" containerID="28dabe457ba3c772d3e1b22e40dc182bb317356ae9a7fb7a66dbfe7373ba5301" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.411498 4989 scope.go:117] "RemoveContainer" containerID="d017868364c6651a0b554c3f50fcbb8eff0538a85118f7409530e13d506b71e4" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.442523 4989 scope.go:117] "RemoveContainer" containerID="d9eaa9f17cb69081351d566ddc882fff1a56a760bb9adde010a3aa60706b6d73" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.466172 4989 scope.go:117] "RemoveContainer" containerID="8ea9658e6471defa08591c862404ca7ebfe7d45bf4349daa637e2559c24e8c21" Oct 06 10:08:20 crc kubenswrapper[4989]: I1006 10:08:20.492468 4989 scope.go:117] "RemoveContainer" containerID="035e02d2fe942373e5d3d40748d687003a90c7fec19f1a6def891a41dfbb8386" Oct 06 10:08:24 crc kubenswrapper[4989]: I1006 10:08:24.936404 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:08:24 crc kubenswrapper[4989]: E1006 10:08:24.937329 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:08:35 crc kubenswrapper[4989]: I1006 10:08:35.936827 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:08:35 crc kubenswrapper[4989]: E1006 10:08:35.938124 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:08:43 crc kubenswrapper[4989]: I1006 10:08:43.571921 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-555c6df57c-dz9gw" Oct 06 10:08:43 crc kubenswrapper[4989]: I1006 10:08:43.600307 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-555c6df57c-dz9gw" podStartSLOduration=32.600287772 podStartE2EDuration="32.600287772s" podCreationTimestamp="2025-10-06 10:08:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:08:13.652392247 +0000 UTC m=+5344.442417827" watchObservedRunningTime="2025-10-06 10:08:43.600287772 +0000 UTC m=+5374.390313372" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.759382 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.761151 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.763766 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.764580 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-kwkvh" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.766931 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.777439 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.817301 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: E1006 10:08:46.818048 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-44rtz openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="ea67281e-7e4d-472e-9589-e46edfff4ad0" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.825491 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.839334 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.841509 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.858978 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.859034 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44rtz\" (UniqueName: \"kubernetes.io/projected/ea67281e-7e4d-472e-9589-e46edfff4ad0-kube-api-access-44rtz\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.859085 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.860222 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.936389 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:08:46 crc kubenswrapper[4989]: E1006 10:08:46.936878 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960065 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjxlz\" (UniqueName: \"kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960153 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960196 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44rtz\" (UniqueName: \"kubernetes.io/projected/ea67281e-7e4d-472e-9589-e46edfff4ad0-kube-api-access-44rtz\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960251 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960294 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.960317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.961802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: E1006 10:08:46.963359 4989 projected.go:194] Error preparing data for projected volume kube-api-access-44rtz for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ea67281e-7e4d-472e-9589-e46edfff4ad0) does not match the UID in record. The object might have been deleted and then recreated Oct 06 10:08:46 crc kubenswrapper[4989]: E1006 10:08:46.963523 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ea67281e-7e4d-472e-9589-e46edfff4ad0-kube-api-access-44rtz podName:ea67281e-7e4d-472e-9589-e46edfff4ad0 nodeName:}" failed. No retries permitted until 2025-10-06 10:08:47.4635072 +0000 UTC m=+5378.253532780 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-44rtz" (UniqueName: "kubernetes.io/projected/ea67281e-7e4d-472e-9589-e46edfff4ad0-kube-api-access-44rtz") pod "openstackclient" (UID: "ea67281e-7e4d-472e-9589-e46edfff4ad0") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (ea67281e-7e4d-472e-9589-e46edfff4ad0) does not match the UID in record. The object might have been deleted and then recreated Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.966966 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret\") pod \"openstackclient\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.972761 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:46 crc kubenswrapper[4989]: I1006 10:08:46.977842 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ea67281e-7e4d-472e-9589-e46edfff4ad0" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.032673 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.063022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.063092 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.063348 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjxlz\" (UniqueName: \"kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.065079 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.069391 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.081558 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjxlz\" (UniqueName: \"kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz\") pod \"openstackclient\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.164392 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret\") pod \"ea67281e-7e4d-472e-9589-e46edfff4ad0\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.164772 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config\") pod \"ea67281e-7e4d-472e-9589-e46edfff4ad0\" (UID: \"ea67281e-7e4d-472e-9589-e46edfff4ad0\") " Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.165146 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44rtz\" (UniqueName: \"kubernetes.io/projected/ea67281e-7e4d-472e-9589-e46edfff4ad0-kube-api-access-44rtz\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.165138 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ea67281e-7e4d-472e-9589-e46edfff4ad0" (UID: "ea67281e-7e4d-472e-9589-e46edfff4ad0"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.165221 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.168435 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ea67281e-7e4d-472e-9589-e46edfff4ad0" (UID: "ea67281e-7e4d-472e-9589-e46edfff4ad0"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.266704 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.266734 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea67281e-7e4d-472e-9589-e46edfff4ad0-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.609382 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.953783 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea67281e-7e4d-472e-9589-e46edfff4ad0" path="/var/lib/kubelet/pods/ea67281e-7e4d-472e-9589-e46edfff4ad0/volumes" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.985071 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.985785 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0cd01b39-add5-482b-9835-7aa6e59c3d8c","Type":"ContainerStarted","Data":"b658573e0eebd0b881bee255bb430656cf19dfb2d63bae4c4197cddca91f2514"} Oct 06 10:08:47 crc kubenswrapper[4989]: I1006 10:08:47.985929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0cd01b39-add5-482b-9835-7aa6e59c3d8c","Type":"ContainerStarted","Data":"89cf5d902534abcbd6c28b163e19421a8002db20c4208f924622816aa70e251e"} Oct 06 10:08:48 crc kubenswrapper[4989]: I1006 10:08:48.005110 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.00509103 podStartE2EDuration="2.00509103s" podCreationTimestamp="2025-10-06 10:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:08:48.002189386 +0000 UTC m=+5378.792214986" watchObservedRunningTime="2025-10-06 10:08:48.00509103 +0000 UTC m=+5378.795116610" Oct 06 10:08:48 crc kubenswrapper[4989]: I1006 10:08:48.008172 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ea67281e-7e4d-472e-9589-e46edfff4ad0" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" Oct 06 10:08:58 crc kubenswrapper[4989]: I1006 10:08:58.936340 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:08:58 crc kubenswrapper[4989]: E1006 10:08:58.937164 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:09:12 crc kubenswrapper[4989]: I1006 10:09:12.936139 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:09:12 crc kubenswrapper[4989]: E1006 10:09:12.937099 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:09:27 crc kubenswrapper[4989]: I1006 10:09:27.936834 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:09:27 crc kubenswrapper[4989]: E1006 10:09:27.937730 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:09:42 crc kubenswrapper[4989]: I1006 10:09:42.936253 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:09:42 crc kubenswrapper[4989]: E1006 10:09:42.936959 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:09:54 crc kubenswrapper[4989]: I1006 10:09:54.936707 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:09:54 crc kubenswrapper[4989]: E1006 10:09:54.937517 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:10:08 crc kubenswrapper[4989]: I1006 10:10:08.936155 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:10:08 crc kubenswrapper[4989]: E1006 10:10:08.937012 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:10:21 crc kubenswrapper[4989]: I1006 10:10:21.936328 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:10:21 crc kubenswrapper[4989]: E1006 10:10:21.937261 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:10:36 crc kubenswrapper[4989]: I1006 10:10:36.936598 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:10:36 crc kubenswrapper[4989]: E1006 10:10:36.937751 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:10:36 crc kubenswrapper[4989]: I1006 10:10:36.945062 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-drvvq"] Oct 06 10:10:36 crc kubenswrapper[4989]: I1006 10:10:36.946089 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:36 crc kubenswrapper[4989]: I1006 10:10:36.957100 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-drvvq"] Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.010476 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6sl7\" (UniqueName: \"kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7\") pod \"barbican-db-create-drvvq\" (UID: \"f7a2cde9-142d-4338-86f4-972c36750be2\") " pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.111744 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6sl7\" (UniqueName: \"kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7\") pod \"barbican-db-create-drvvq\" (UID: \"f7a2cde9-142d-4338-86f4-972c36750be2\") " pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.131623 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6sl7\" (UniqueName: \"kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7\") pod \"barbican-db-create-drvvq\" (UID: \"f7a2cde9-142d-4338-86f4-972c36750be2\") " pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.265999 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.745951 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-drvvq"] Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.988406 4989 generic.go:334] "Generic (PLEG): container finished" podID="f7a2cde9-142d-4338-86f4-972c36750be2" containerID="cf768e88a3abfe99608b401828618af7036ef4b20ba2055e0b8522e189489eb9" exitCode=0 Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.988444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-drvvq" event={"ID":"f7a2cde9-142d-4338-86f4-972c36750be2","Type":"ContainerDied","Data":"cf768e88a3abfe99608b401828618af7036ef4b20ba2055e0b8522e189489eb9"} Oct 06 10:10:37 crc kubenswrapper[4989]: I1006 10:10:37.988472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-drvvq" event={"ID":"f7a2cde9-142d-4338-86f4-972c36750be2","Type":"ContainerStarted","Data":"5776042a09551c9effa0e04fd06d275996db04c54c6fd1cefbc40aa159b7e0f6"} Oct 06 10:10:39 crc kubenswrapper[4989]: I1006 10:10:39.340388 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:39 crc kubenswrapper[4989]: I1006 10:10:39.454175 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6sl7\" (UniqueName: \"kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7\") pod \"f7a2cde9-142d-4338-86f4-972c36750be2\" (UID: \"f7a2cde9-142d-4338-86f4-972c36750be2\") " Oct 06 10:10:39 crc kubenswrapper[4989]: I1006 10:10:39.464360 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7" (OuterVolumeSpecName: "kube-api-access-v6sl7") pod "f7a2cde9-142d-4338-86f4-972c36750be2" (UID: "f7a2cde9-142d-4338-86f4-972c36750be2"). InnerVolumeSpecName "kube-api-access-v6sl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:39 crc kubenswrapper[4989]: I1006 10:10:39.557082 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6sl7\" (UniqueName: \"kubernetes.io/projected/f7a2cde9-142d-4338-86f4-972c36750be2-kube-api-access-v6sl7\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:40 crc kubenswrapper[4989]: I1006 10:10:40.013635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-drvvq" event={"ID":"f7a2cde9-142d-4338-86f4-972c36750be2","Type":"ContainerDied","Data":"5776042a09551c9effa0e04fd06d275996db04c54c6fd1cefbc40aa159b7e0f6"} Oct 06 10:10:40 crc kubenswrapper[4989]: I1006 10:10:40.013744 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5776042a09551c9effa0e04fd06d275996db04c54c6fd1cefbc40aa159b7e0f6" Oct 06 10:10:40 crc kubenswrapper[4989]: I1006 10:10:40.013824 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-drvvq" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.970313 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-9766-account-create-5lr7p"] Oct 06 10:10:46 crc kubenswrapper[4989]: E1006 10:10:46.971920 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a2cde9-142d-4338-86f4-972c36750be2" containerName="mariadb-database-create" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.971939 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a2cde9-142d-4338-86f4-972c36750be2" containerName="mariadb-database-create" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.972149 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a2cde9-142d-4338-86f4-972c36750be2" containerName="mariadb-database-create" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.973207 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.975991 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 06 10:10:46 crc kubenswrapper[4989]: I1006 10:10:46.984586 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9766-account-create-5lr7p"] Oct 06 10:10:47 crc kubenswrapper[4989]: I1006 10:10:47.097482 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkdm9\" (UniqueName: \"kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9\") pod \"barbican-9766-account-create-5lr7p\" (UID: \"3f17f9bc-804b-4244-bb7d-b297e1069061\") " pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:47 crc kubenswrapper[4989]: I1006 10:10:47.199603 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkdm9\" (UniqueName: \"kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9\") pod \"barbican-9766-account-create-5lr7p\" (UID: \"3f17f9bc-804b-4244-bb7d-b297e1069061\") " pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:47 crc kubenswrapper[4989]: I1006 10:10:47.224787 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkdm9\" (UniqueName: \"kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9\") pod \"barbican-9766-account-create-5lr7p\" (UID: \"3f17f9bc-804b-4244-bb7d-b297e1069061\") " pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:47 crc kubenswrapper[4989]: I1006 10:10:47.307151 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:47 crc kubenswrapper[4989]: I1006 10:10:47.710267 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9766-account-create-5lr7p"] Oct 06 10:10:47 crc kubenswrapper[4989]: W1006 10:10:47.715216 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f17f9bc_804b_4244_bb7d_b297e1069061.slice/crio-fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa WatchSource:0}: Error finding container fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa: Status 404 returned error can't find the container with id fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa Oct 06 10:10:48 crc kubenswrapper[4989]: I1006 10:10:48.088584 4989 generic.go:334] "Generic (PLEG): container finished" podID="3f17f9bc-804b-4244-bb7d-b297e1069061" containerID="a3332bed5eb24cbc9d50b7e71dc1715e26fc0e8d875e96f3553f84a4cac7ee04" exitCode=0 Oct 06 10:10:48 crc kubenswrapper[4989]: I1006 10:10:48.088643 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9766-account-create-5lr7p" event={"ID":"3f17f9bc-804b-4244-bb7d-b297e1069061","Type":"ContainerDied","Data":"a3332bed5eb24cbc9d50b7e71dc1715e26fc0e8d875e96f3553f84a4cac7ee04"} Oct 06 10:10:48 crc kubenswrapper[4989]: I1006 10:10:48.088986 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9766-account-create-5lr7p" event={"ID":"3f17f9bc-804b-4244-bb7d-b297e1069061","Type":"ContainerStarted","Data":"fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa"} Oct 06 10:10:49 crc kubenswrapper[4989]: I1006 10:10:49.414957 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:49 crc kubenswrapper[4989]: I1006 10:10:49.537240 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkdm9\" (UniqueName: \"kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9\") pod \"3f17f9bc-804b-4244-bb7d-b297e1069061\" (UID: \"3f17f9bc-804b-4244-bb7d-b297e1069061\") " Oct 06 10:10:49 crc kubenswrapper[4989]: I1006 10:10:49.543153 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9" (OuterVolumeSpecName: "kube-api-access-jkdm9") pod "3f17f9bc-804b-4244-bb7d-b297e1069061" (UID: "3f17f9bc-804b-4244-bb7d-b297e1069061"). InnerVolumeSpecName "kube-api-access-jkdm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:49 crc kubenswrapper[4989]: I1006 10:10:49.639784 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkdm9\" (UniqueName: \"kubernetes.io/projected/3f17f9bc-804b-4244-bb7d-b297e1069061-kube-api-access-jkdm9\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:50 crc kubenswrapper[4989]: I1006 10:10:50.109234 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9766-account-create-5lr7p" event={"ID":"3f17f9bc-804b-4244-bb7d-b297e1069061","Type":"ContainerDied","Data":"fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa"} Oct 06 10:10:50 crc kubenswrapper[4989]: I1006 10:10:50.109308 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb88e057891d19ab08ad3eefa13c85c46e057147e74ec374f3e9cc5649ae81aa" Oct 06 10:10:50 crc kubenswrapper[4989]: I1006 10:10:50.109359 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9766-account-create-5lr7p" Oct 06 10:10:51 crc kubenswrapper[4989]: I1006 10:10:51.936802 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:10:51 crc kubenswrapper[4989]: E1006 10:10:51.937370 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.254069 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-jhgpb"] Oct 06 10:10:52 crc kubenswrapper[4989]: E1006 10:10:52.254502 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f17f9bc-804b-4244-bb7d-b297e1069061" containerName="mariadb-account-create" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.254522 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f17f9bc-804b-4244-bb7d-b297e1069061" containerName="mariadb-account-create" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.254736 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f17f9bc-804b-4244-bb7d-b297e1069061" containerName="mariadb-account-create" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.255384 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.257731 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.258210 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2lw6t" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.302018 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jhgpb"] Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.386612 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9vv5\" (UniqueName: \"kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.386679 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.386738 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.488670 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9vv5\" (UniqueName: \"kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.488737 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.488812 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.494115 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.495043 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.518137 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9vv5\" (UniqueName: \"kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5\") pod \"barbican-db-sync-jhgpb\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:52 crc kubenswrapper[4989]: I1006 10:10:52.582816 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:53 crc kubenswrapper[4989]: I1006 10:10:53.139352 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-jhgpb"] Oct 06 10:10:54 crc kubenswrapper[4989]: I1006 10:10:54.140061 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jhgpb" event={"ID":"6d8844c1-5796-4d53-a114-856435a3d743","Type":"ContainerStarted","Data":"cb45ef9cf20362313e0a31a718e8e7f6491936616a109088467309489e91655d"} Oct 06 10:10:54 crc kubenswrapper[4989]: I1006 10:10:54.140372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jhgpb" event={"ID":"6d8844c1-5796-4d53-a114-856435a3d743","Type":"ContainerStarted","Data":"9e2293d251b96e2a6466a16266b97782fdfee94b2b40f26c5b42f2faee6ff9d7"} Oct 06 10:10:54 crc kubenswrapper[4989]: I1006 10:10:54.159078 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-jhgpb" podStartSLOduration=2.159056025 podStartE2EDuration="2.159056025s" podCreationTimestamp="2025-10-06 10:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:54.154523005 +0000 UTC m=+5504.944548595" watchObservedRunningTime="2025-10-06 10:10:54.159056025 +0000 UTC m=+5504.949081625" Oct 06 10:10:55 crc kubenswrapper[4989]: I1006 10:10:55.150463 4989 generic.go:334] "Generic (PLEG): container finished" podID="6d8844c1-5796-4d53-a114-856435a3d743" containerID="cb45ef9cf20362313e0a31a718e8e7f6491936616a109088467309489e91655d" exitCode=0 Oct 06 10:10:55 crc kubenswrapper[4989]: I1006 10:10:55.150508 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jhgpb" event={"ID":"6d8844c1-5796-4d53-a114-856435a3d743","Type":"ContainerDied","Data":"cb45ef9cf20362313e0a31a718e8e7f6491936616a109088467309489e91655d"} Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.472832 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.581945 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data\") pod \"6d8844c1-5796-4d53-a114-856435a3d743\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.581986 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle\") pod \"6d8844c1-5796-4d53-a114-856435a3d743\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.582056 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9vv5\" (UniqueName: \"kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5\") pod \"6d8844c1-5796-4d53-a114-856435a3d743\" (UID: \"6d8844c1-5796-4d53-a114-856435a3d743\") " Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.587982 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5" (OuterVolumeSpecName: "kube-api-access-v9vv5") pod "6d8844c1-5796-4d53-a114-856435a3d743" (UID: "6d8844c1-5796-4d53-a114-856435a3d743"). InnerVolumeSpecName "kube-api-access-v9vv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.588238 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6d8844c1-5796-4d53-a114-856435a3d743" (UID: "6d8844c1-5796-4d53-a114-856435a3d743"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.605409 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d8844c1-5796-4d53-a114-856435a3d743" (UID: "6d8844c1-5796-4d53-a114-856435a3d743"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.684161 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9vv5\" (UniqueName: \"kubernetes.io/projected/6d8844c1-5796-4d53-a114-856435a3d743-kube-api-access-v9vv5\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.684190 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:56 crc kubenswrapper[4989]: I1006 10:10:56.684200 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8844c1-5796-4d53-a114-856435a3d743-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.172872 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-jhgpb" event={"ID":"6d8844c1-5796-4d53-a114-856435a3d743","Type":"ContainerDied","Data":"9e2293d251b96e2a6466a16266b97782fdfee94b2b40f26c5b42f2faee6ff9d7"} Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.172960 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e2293d251b96e2a6466a16266b97782fdfee94b2b40f26c5b42f2faee6ff9d7" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.173068 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-jhgpb" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.368600 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7f4d8b446c-sdh4z"] Oct 06 10:10:57 crc kubenswrapper[4989]: E1006 10:10:57.369127 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8844c1-5796-4d53-a114-856435a3d743" containerName="barbican-db-sync" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.369155 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8844c1-5796-4d53-a114-856435a3d743" containerName="barbican-db-sync" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.369382 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d8844c1-5796-4d53-a114-856435a3d743" containerName="barbican-db-sync" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.370491 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.376706 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6bf65bc98-6gs8k"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.377182 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2lw6t" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.377343 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.377613 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.377978 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.380530 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.385500 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f4d8b446c-sdh4z"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.402199 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bf65bc98-6gs8k"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.471854 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.474076 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.479844 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.511929 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512018 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data-custom\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512117 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm56m\" (UniqueName: \"kubernetes.io/projected/08847650-e4e2-465b-ba93-bab730da6f00-kube-api-access-qm56m\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512164 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512198 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl8p2\" (UniqueName: \"kubernetes.io/projected/af19d793-7709-494b-a5de-6696bc6ab72e-kube-api-access-sl8p2\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512498 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08847650-e4e2-465b-ba93-bab730da6f00-logs\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512636 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data-custom\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512695 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af19d793-7709-494b-a5de-6696bc6ab72e-logs\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.512759 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-combined-ca-bundle\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.560146 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-65988db468-g5rf5"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.561812 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.564964 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.583378 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65988db468-g5rf5"] Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-combined-ca-bundle\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614250 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614274 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data-custom\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614301 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614339 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm56m\" (UniqueName: \"kubernetes.io/projected/08847650-e4e2-465b-ba93-bab730da6f00-kube-api-access-qm56m\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614366 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614410 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl8p2\" (UniqueName: \"kubernetes.io/projected/af19d793-7709-494b-a5de-6696bc6ab72e-kube-api-access-sl8p2\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614436 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614634 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614694 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08847650-e4e2-465b-ba93-bab730da6f00-logs\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614751 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data-custom\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614775 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af19d793-7709-494b-a5de-6696bc6ab72e-logs\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.614798 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss9vr\" (UniqueName: \"kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.615619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08847650-e4e2-465b-ba93-bab730da6f00-logs\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.619600 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.620632 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af19d793-7709-494b-a5de-6696bc6ab72e-logs\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.621106 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-config-data-custom\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.622588 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-combined-ca-bundle\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.624277 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data-custom\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.625224 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af19d793-7709-494b-a5de-6696bc6ab72e-config-data\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.629501 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08847650-e4e2-465b-ba93-bab730da6f00-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.633783 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl8p2\" (UniqueName: \"kubernetes.io/projected/af19d793-7709-494b-a5de-6696bc6ab72e-kube-api-access-sl8p2\") pod \"barbican-worker-7f4d8b446c-sdh4z\" (UID: \"af19d793-7709-494b-a5de-6696bc6ab72e\") " pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.636531 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm56m\" (UniqueName: \"kubernetes.io/projected/08847650-e4e2-465b-ba93-bab730da6f00-kube-api-access-qm56m\") pod \"barbican-keystone-listener-6bf65bc98-6gs8k\" (UID: \"08847650-e4e2-465b-ba93-bab730da6f00\") " pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.697236 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716223 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716673 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-combined-ca-bundle\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716723 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data-custom\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716755 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xczdw\" (UniqueName: \"kubernetes.io/projected/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-kube-api-access-xczdw\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716794 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss9vr\" (UniqueName: \"kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716840 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716886 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716913 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.716961 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.717081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-logs\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.717129 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.717879 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.717961 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.718028 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.719297 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.736084 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss9vr\" (UniqueName: \"kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr\") pod \"dnsmasq-dns-774d46d7f7-bmc7j\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.819604 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820245 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-logs\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820382 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-combined-ca-bundle\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820402 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data-custom\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820425 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xczdw\" (UniqueName: \"kubernetes.io/projected/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-kube-api-access-xczdw\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.820901 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-logs\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.829578 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-combined-ca-bundle\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.830505 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.832113 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-config-data-custom\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.837586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xczdw\" (UniqueName: \"kubernetes.io/projected/10c3c6f7-811a-40c0-92c2-63d02b2ecd2b-kube-api-access-xczdw\") pod \"barbican-api-65988db468-g5rf5\" (UID: \"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b\") " pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:57 crc kubenswrapper[4989]: I1006 10:10:57.882718 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:58 crc kubenswrapper[4989]: I1006 10:10:58.200643 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7f4d8b446c-sdh4z"] Oct 06 10:10:58 crc kubenswrapper[4989]: I1006 10:10:58.294979 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bf65bc98-6gs8k"] Oct 06 10:10:58 crc kubenswrapper[4989]: I1006 10:10:58.450115 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:10:58 crc kubenswrapper[4989]: I1006 10:10:58.478921 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-65988db468-g5rf5"] Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.193250 4989 generic.go:334] "Generic (PLEG): container finished" podID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerID="bdbbd9375919c80d764b03c9b8cc82a9429e5914a0294927909d2cf84261e0bd" exitCode=0 Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.193361 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" event={"ID":"7013d8e4-6361-41bf-8b32-82d3a2b7e557","Type":"ContainerDied","Data":"bdbbd9375919c80d764b03c9b8cc82a9429e5914a0294927909d2cf84261e0bd"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.193682 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" event={"ID":"7013d8e4-6361-41bf-8b32-82d3a2b7e557","Type":"ContainerStarted","Data":"0df2b0e870e3eaf7465ed8524c0ccbf11ae3f59249c747de404ed4299eda7836"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.198105 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" event={"ID":"08847650-e4e2-465b-ba93-bab730da6f00","Type":"ContainerStarted","Data":"5093f931d07ac21e8ba5f7b9f68041c2461f26e53b9ad645e2e0bd19e844002c"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.198153 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" event={"ID":"08847650-e4e2-465b-ba93-bab730da6f00","Type":"ContainerStarted","Data":"513c0e0e1715202b2f26337b8f75907c1c29b25f5ea52b10617b28c4759f642f"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.198164 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" event={"ID":"08847650-e4e2-465b-ba93-bab730da6f00","Type":"ContainerStarted","Data":"1aeefdcdba12a6adf5d7282468870b39ad1389f7f266c01ffd94d561f7a1b8e5"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.207184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" event={"ID":"af19d793-7709-494b-a5de-6696bc6ab72e","Type":"ContainerStarted","Data":"68d867f8b5500858f5a65bcb9d6ed3e16ec2c6bc86723aa1ba040b1550c59329"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.207232 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" event={"ID":"af19d793-7709-494b-a5de-6696bc6ab72e","Type":"ContainerStarted","Data":"eeaf8bf679b6f65f116b07090a02de59c93f4054de5f8c76d92973d5f77f44ea"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.207241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" event={"ID":"af19d793-7709-494b-a5de-6696bc6ab72e","Type":"ContainerStarted","Data":"223d0d5e960aaf879c6d6082b52bcbce4fa3719fcdf5c64cae144a829c3d306d"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.219306 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65988db468-g5rf5" event={"ID":"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b","Type":"ContainerStarted","Data":"0025af896dd1093b11f3525079dd53ffcd5e676fc9f4816a6ab62f72a119f9d8"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.219361 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65988db468-g5rf5" event={"ID":"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b","Type":"ContainerStarted","Data":"1151d72dead5216dbedfe8478612632a6b28a26adeac9849cbf594474fc678e9"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.219378 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-65988db468-g5rf5" event={"ID":"10c3c6f7-811a-40c0-92c2-63d02b2ecd2b","Type":"ContainerStarted","Data":"9d0c28af6d79ca9d3c6ba2c411850e2b274f9fdeeb4bcc88e6abf0845117f336"} Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.220330 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.220368 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.239871 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7f4d8b446c-sdh4z" podStartSLOduration=2.239851891 podStartE2EDuration="2.239851891s" podCreationTimestamp="2025-10-06 10:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:59.236076872 +0000 UTC m=+5510.026102462" watchObservedRunningTime="2025-10-06 10:10:59.239851891 +0000 UTC m=+5510.029877471" Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.254517 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6bf65bc98-6gs8k" podStartSLOduration=2.254498132 podStartE2EDuration="2.254498132s" podCreationTimestamp="2025-10-06 10:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:59.253228146 +0000 UTC m=+5510.043253726" watchObservedRunningTime="2025-10-06 10:10:59.254498132 +0000 UTC m=+5510.044523712" Oct 06 10:10:59 crc kubenswrapper[4989]: I1006 10:10:59.282561 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-65988db468-g5rf5" podStartSLOduration=2.28254112 podStartE2EDuration="2.28254112s" podCreationTimestamp="2025-10-06 10:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:10:59.269223516 +0000 UTC m=+5510.059249096" watchObservedRunningTime="2025-10-06 10:10:59.28254112 +0000 UTC m=+5510.072566700" Oct 06 10:11:00 crc kubenswrapper[4989]: I1006 10:11:00.230444 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" event={"ID":"7013d8e4-6361-41bf-8b32-82d3a2b7e557","Type":"ContainerStarted","Data":"5e7d035e7a8d78246173cd60a429bf923ae69673a36dbad70fc327a5bb20ffaa"} Oct 06 10:11:00 crc kubenswrapper[4989]: I1006 10:11:00.254758 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" podStartSLOduration=3.254742103 podStartE2EDuration="3.254742103s" podCreationTimestamp="2025-10-06 10:10:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:00.253126236 +0000 UTC m=+5511.043151826" watchObservedRunningTime="2025-10-06 10:11:00.254742103 +0000 UTC m=+5511.044767683" Oct 06 10:11:01 crc kubenswrapper[4989]: I1006 10:11:01.241043 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:11:05 crc kubenswrapper[4989]: I1006 10:11:05.935535 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:11:05 crc kubenswrapper[4989]: E1006 10:11:05.936489 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:11:07 crc kubenswrapper[4989]: I1006 10:11:07.821854 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:11:07 crc kubenswrapper[4989]: I1006 10:11:07.895067 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:11:07 crc kubenswrapper[4989]: I1006 10:11:07.895322 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="dnsmasq-dns" containerID="cri-o://891fa4f95e00f2312f6f291e60e13aeb9696640b2d2587c1a29020f62707e947" gracePeriod=10 Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.003830 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.15:5353: connect: connection refused" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.335218 4989 generic.go:334] "Generic (PLEG): container finished" podID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerID="891fa4f95e00f2312f6f291e60e13aeb9696640b2d2587c1a29020f62707e947" exitCode=0 Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.335432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" event={"ID":"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a","Type":"ContainerDied","Data":"891fa4f95e00f2312f6f291e60e13aeb9696640b2d2587c1a29020f62707e947"} Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.421797 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.529356 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb\") pod \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.529402 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config\") pod \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.529604 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc\") pod \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.529708 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb\") pod \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.529739 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgg75\" (UniqueName: \"kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75\") pod \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\" (UID: \"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a\") " Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.548854 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75" (OuterVolumeSpecName: "kube-api-access-dgg75") pod "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" (UID: "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a"). InnerVolumeSpecName "kube-api-access-dgg75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.576293 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" (UID: "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.579622 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" (UID: "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.580616 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config" (OuterVolumeSpecName: "config") pod "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" (UID: "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.581367 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" (UID: "5b29f548-7a90-4af9-a9c8-fb4d6e8c905a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.632527 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.632582 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgg75\" (UniqueName: \"kubernetes.io/projected/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-kube-api-access-dgg75\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.632597 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.632612 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:08 crc kubenswrapper[4989]: I1006 10:11:08.632624 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.336132 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.354635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" event={"ID":"5b29f548-7a90-4af9-a9c8-fb4d6e8c905a","Type":"ContainerDied","Data":"26a0ce4cf87509c345da12e1b6cf44ceaca1b67decd05ad310c8f1454e628ca3"} Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.355099 4989 scope.go:117] "RemoveContainer" containerID="891fa4f95e00f2312f6f291e60e13aeb9696640b2d2587c1a29020f62707e947" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.355318 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d4b57f6f-j7r5m" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.384381 4989 scope.go:117] "RemoveContainer" containerID="c1e5b7c39d73b84bac99e546e3a04747c71ba8feca977ca16e28ab9ef7ac316e" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.401441 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.412331 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85d4b57f6f-j7r5m"] Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.634275 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-65988db468-g5rf5" Oct 06 10:11:09 crc kubenswrapper[4989]: I1006 10:11:09.948131 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" path="/var/lib/kubelet/pods/5b29f548-7a90-4af9-a9c8-fb4d6e8c905a/volumes" Oct 06 10:11:17 crc kubenswrapper[4989]: I1006 10:11:17.936319 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:11:17 crc kubenswrapper[4989]: E1006 10:11:17.937167 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.422460 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-5j6jz"] Oct 06 10:11:24 crc kubenswrapper[4989]: E1006 10:11:24.423465 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="init" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.423484 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="init" Oct 06 10:11:24 crc kubenswrapper[4989]: E1006 10:11:24.423513 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="dnsmasq-dns" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.423522 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="dnsmasq-dns" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.423776 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b29f548-7a90-4af9-a9c8-fb4d6e8c905a" containerName="dnsmasq-dns" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.424486 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.442900 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5j6jz"] Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.544223 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdtxt\" (UniqueName: \"kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt\") pod \"neutron-db-create-5j6jz\" (UID: \"87efa56a-fb99-42e8-91d6-680def8278ed\") " pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.645700 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdtxt\" (UniqueName: \"kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt\") pod \"neutron-db-create-5j6jz\" (UID: \"87efa56a-fb99-42e8-91d6-680def8278ed\") " pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.668559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdtxt\" (UniqueName: \"kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt\") pod \"neutron-db-create-5j6jz\" (UID: \"87efa56a-fb99-42e8-91d6-680def8278ed\") " pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:24 crc kubenswrapper[4989]: I1006 10:11:24.750107 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:25 crc kubenswrapper[4989]: I1006 10:11:25.199523 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5j6jz"] Oct 06 10:11:25 crc kubenswrapper[4989]: I1006 10:11:25.513563 4989 generic.go:334] "Generic (PLEG): container finished" podID="87efa56a-fb99-42e8-91d6-680def8278ed" containerID="7a2ff0e4f886535744336ed8e4c102b8504eff528a009a998f52025f453c7ddf" exitCode=0 Oct 06 10:11:25 crc kubenswrapper[4989]: I1006 10:11:25.513647 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5j6jz" event={"ID":"87efa56a-fb99-42e8-91d6-680def8278ed","Type":"ContainerDied","Data":"7a2ff0e4f886535744336ed8e4c102b8504eff528a009a998f52025f453c7ddf"} Oct 06 10:11:25 crc kubenswrapper[4989]: I1006 10:11:25.514097 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5j6jz" event={"ID":"87efa56a-fb99-42e8-91d6-680def8278ed","Type":"ContainerStarted","Data":"a48c08343c210ec164461ebc12eb15712fff02423ea433cfa4a35ab5279dd906"} Oct 06 10:11:26 crc kubenswrapper[4989]: I1006 10:11:26.964901 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.089824 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdtxt\" (UniqueName: \"kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt\") pod \"87efa56a-fb99-42e8-91d6-680def8278ed\" (UID: \"87efa56a-fb99-42e8-91d6-680def8278ed\") " Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.095849 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt" (OuterVolumeSpecName: "kube-api-access-jdtxt") pod "87efa56a-fb99-42e8-91d6-680def8278ed" (UID: "87efa56a-fb99-42e8-91d6-680def8278ed"). InnerVolumeSpecName "kube-api-access-jdtxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.192604 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdtxt\" (UniqueName: \"kubernetes.io/projected/87efa56a-fb99-42e8-91d6-680def8278ed-kube-api-access-jdtxt\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.537705 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5j6jz" event={"ID":"87efa56a-fb99-42e8-91d6-680def8278ed","Type":"ContainerDied","Data":"a48c08343c210ec164461ebc12eb15712fff02423ea433cfa4a35ab5279dd906"} Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.537751 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5j6jz" Oct 06 10:11:27 crc kubenswrapper[4989]: I1006 10:11:27.537756 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48c08343c210ec164461ebc12eb15712fff02423ea433cfa4a35ab5279dd906" Oct 06 10:11:32 crc kubenswrapper[4989]: I1006 10:11:32.936367 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:11:32 crc kubenswrapper[4989]: E1006 10:11:32.937117 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.514273 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86e0-account-create-rc22c"] Oct 06 10:11:34 crc kubenswrapper[4989]: E1006 10:11:34.516373 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87efa56a-fb99-42e8-91d6-680def8278ed" containerName="mariadb-database-create" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.516505 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="87efa56a-fb99-42e8-91d6-680def8278ed" containerName="mariadb-database-create" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.518248 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="87efa56a-fb99-42e8-91d6-680def8278ed" containerName="mariadb-database-create" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.519315 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.522034 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.527777 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86e0-account-create-rc22c"] Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.631581 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vclwj\" (UniqueName: \"kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj\") pod \"neutron-86e0-account-create-rc22c\" (UID: \"5646c4e1-673f-4571-9d34-9558a1d7e0ce\") " pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.733679 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vclwj\" (UniqueName: \"kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj\") pod \"neutron-86e0-account-create-rc22c\" (UID: \"5646c4e1-673f-4571-9d34-9558a1d7e0ce\") " pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.757793 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vclwj\" (UniqueName: \"kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj\") pod \"neutron-86e0-account-create-rc22c\" (UID: \"5646c4e1-673f-4571-9d34-9558a1d7e0ce\") " pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:34 crc kubenswrapper[4989]: I1006 10:11:34.881949 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:35 crc kubenswrapper[4989]: I1006 10:11:35.328613 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86e0-account-create-rc22c"] Oct 06 10:11:35 crc kubenswrapper[4989]: I1006 10:11:35.628800 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86e0-account-create-rc22c" event={"ID":"5646c4e1-673f-4571-9d34-9558a1d7e0ce","Type":"ContainerStarted","Data":"6fe6755f0bb0e9cc7ef1281e2d66bc095e6147a885eb3c3aba97bcb22037db23"} Oct 06 10:11:35 crc kubenswrapper[4989]: I1006 10:11:35.629111 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86e0-account-create-rc22c" event={"ID":"5646c4e1-673f-4571-9d34-9558a1d7e0ce","Type":"ContainerStarted","Data":"963e63b6202287271369017424264f4d3b94b1774c9f199a021230c3e31bbf07"} Oct 06 10:11:35 crc kubenswrapper[4989]: I1006 10:11:35.658468 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86e0-account-create-rc22c" podStartSLOduration=1.658449254 podStartE2EDuration="1.658449254s" podCreationTimestamp="2025-10-06 10:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:35.650351861 +0000 UTC m=+5546.440377451" watchObservedRunningTime="2025-10-06 10:11:35.658449254 +0000 UTC m=+5546.448474834" Oct 06 10:11:36 crc kubenswrapper[4989]: I1006 10:11:36.641380 4989 generic.go:334] "Generic (PLEG): container finished" podID="5646c4e1-673f-4571-9d34-9558a1d7e0ce" containerID="6fe6755f0bb0e9cc7ef1281e2d66bc095e6147a885eb3c3aba97bcb22037db23" exitCode=0 Oct 06 10:11:36 crc kubenswrapper[4989]: I1006 10:11:36.641440 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86e0-account-create-rc22c" event={"ID":"5646c4e1-673f-4571-9d34-9558a1d7e0ce","Type":"ContainerDied","Data":"6fe6755f0bb0e9cc7ef1281e2d66bc095e6147a885eb3c3aba97bcb22037db23"} Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.002041 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.090433 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vclwj\" (UniqueName: \"kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj\") pod \"5646c4e1-673f-4571-9d34-9558a1d7e0ce\" (UID: \"5646c4e1-673f-4571-9d34-9558a1d7e0ce\") " Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.096866 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj" (OuterVolumeSpecName: "kube-api-access-vclwj") pod "5646c4e1-673f-4571-9d34-9558a1d7e0ce" (UID: "5646c4e1-673f-4571-9d34-9558a1d7e0ce"). InnerVolumeSpecName "kube-api-access-vclwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.192294 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vclwj\" (UniqueName: \"kubernetes.io/projected/5646c4e1-673f-4571-9d34-9558a1d7e0ce-kube-api-access-vclwj\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.676469 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86e0-account-create-rc22c" event={"ID":"5646c4e1-673f-4571-9d34-9558a1d7e0ce","Type":"ContainerDied","Data":"963e63b6202287271369017424264f4d3b94b1774c9f199a021230c3e31bbf07"} Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.676517 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="963e63b6202287271369017424264f4d3b94b1774c9f199a021230c3e31bbf07" Oct 06 10:11:38 crc kubenswrapper[4989]: I1006 10:11:38.676591 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86e0-account-create-rc22c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.279938 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:39 crc kubenswrapper[4989]: E1006 10:11:39.280570 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5646c4e1-673f-4571-9d34-9558a1d7e0ce" containerName="mariadb-account-create" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.280599 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5646c4e1-673f-4571-9d34-9558a1d7e0ce" containerName="mariadb-account-create" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.281073 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5646c4e1-673f-4571-9d34-9558a1d7e0ce" containerName="mariadb-account-create" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.283481 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.293063 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.413819 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtrdp\" (UniqueName: \"kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.413897 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.413968 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.515793 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.516092 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtrdp\" (UniqueName: \"kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.516181 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.516614 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.516723 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.552594 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtrdp\" (UniqueName: \"kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp\") pod \"redhat-operators-gtx8c\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.615358 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.763304 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lhdlc"] Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.766957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.769912 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qbb8h" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.771469 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.771719 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.787392 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lhdlc"] Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.824931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swqn4\" (UniqueName: \"kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.825036 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.828872 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.930407 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.930889 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.930949 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swqn4\" (UniqueName: \"kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.942506 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.954512 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:39 crc kubenswrapper[4989]: I1006 10:11:39.972302 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swqn4\" (UniqueName: \"kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4\") pod \"neutron-db-sync-lhdlc\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.100967 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.146162 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:40 crc kubenswrapper[4989]: W1006 10:11:40.158926 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8878b926_e5c2_4c44_8858_615fa2aeb92f.slice/crio-dfe088b480547bb8e53a0e407a68d73f0fb61d454f1e8892db0e36b82dd50164 WatchSource:0}: Error finding container dfe088b480547bb8e53a0e407a68d73f0fb61d454f1e8892db0e36b82dd50164: Status 404 returned error can't find the container with id dfe088b480547bb8e53a0e407a68d73f0fb61d454f1e8892db0e36b82dd50164 Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.569595 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lhdlc"] Oct 06 10:11:40 crc kubenswrapper[4989]: W1006 10:11:40.583920 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod678f61cd_297a_469e_b897_ec94ec29daf0.slice/crio-4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365 WatchSource:0}: Error finding container 4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365: Status 404 returned error can't find the container with id 4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365 Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.698048 4989 generic.go:334] "Generic (PLEG): container finished" podID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerID="2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8" exitCode=0 Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.698092 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerDied","Data":"2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8"} Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.698137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerStarted","Data":"dfe088b480547bb8e53a0e407a68d73f0fb61d454f1e8892db0e36b82dd50164"} Oct 06 10:11:40 crc kubenswrapper[4989]: I1006 10:11:40.699914 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lhdlc" event={"ID":"678f61cd-297a-469e-b897-ec94ec29daf0","Type":"ContainerStarted","Data":"4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365"} Oct 06 10:11:41 crc kubenswrapper[4989]: I1006 10:11:41.719579 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lhdlc" event={"ID":"678f61cd-297a-469e-b897-ec94ec29daf0","Type":"ContainerStarted","Data":"423056df6e55c2701a50b591bc8a5733a660ec921c66153e04b23600d8f28358"} Oct 06 10:11:41 crc kubenswrapper[4989]: I1006 10:11:41.723929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerStarted","Data":"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222"} Oct 06 10:11:41 crc kubenswrapper[4989]: I1006 10:11:41.745234 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lhdlc" podStartSLOduration=2.745208896 podStartE2EDuration="2.745208896s" podCreationTimestamp="2025-10-06 10:11:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:41.7401331 +0000 UTC m=+5552.530158720" watchObservedRunningTime="2025-10-06 10:11:41.745208896 +0000 UTC m=+5552.535234516" Oct 06 10:11:42 crc kubenswrapper[4989]: I1006 10:11:42.734327 4989 generic.go:334] "Generic (PLEG): container finished" podID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerID="5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222" exitCode=0 Oct 06 10:11:42 crc kubenswrapper[4989]: I1006 10:11:42.734520 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerDied","Data":"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222"} Oct 06 10:11:43 crc kubenswrapper[4989]: I1006 10:11:43.746182 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerStarted","Data":"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e"} Oct 06 10:11:43 crc kubenswrapper[4989]: I1006 10:11:43.764709 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gtx8c" podStartSLOduration=2.233060214 podStartE2EDuration="4.764687794s" podCreationTimestamp="2025-10-06 10:11:39 +0000 UTC" firstStartedPulling="2025-10-06 10:11:40.69964909 +0000 UTC m=+5551.489674670" lastFinishedPulling="2025-10-06 10:11:43.23127667 +0000 UTC m=+5554.021302250" observedRunningTime="2025-10-06 10:11:43.764160849 +0000 UTC m=+5554.554186429" watchObservedRunningTime="2025-10-06 10:11:43.764687794 +0000 UTC m=+5554.554713384" Oct 06 10:11:43 crc kubenswrapper[4989]: I1006 10:11:43.936996 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:11:43 crc kubenswrapper[4989]: E1006 10:11:43.937247 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:11:45 crc kubenswrapper[4989]: I1006 10:11:45.765716 4989 generic.go:334] "Generic (PLEG): container finished" podID="678f61cd-297a-469e-b897-ec94ec29daf0" containerID="423056df6e55c2701a50b591bc8a5733a660ec921c66153e04b23600d8f28358" exitCode=0 Oct 06 10:11:45 crc kubenswrapper[4989]: I1006 10:11:45.765791 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lhdlc" event={"ID":"678f61cd-297a-469e-b897-ec94ec29daf0","Type":"ContainerDied","Data":"423056df6e55c2701a50b591bc8a5733a660ec921c66153e04b23600d8f28358"} Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.097040 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.170921 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config\") pod \"678f61cd-297a-469e-b897-ec94ec29daf0\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.170979 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle\") pod \"678f61cd-297a-469e-b897-ec94ec29daf0\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.171016 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swqn4\" (UniqueName: \"kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4\") pod \"678f61cd-297a-469e-b897-ec94ec29daf0\" (UID: \"678f61cd-297a-469e-b897-ec94ec29daf0\") " Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.175635 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4" (OuterVolumeSpecName: "kube-api-access-swqn4") pod "678f61cd-297a-469e-b897-ec94ec29daf0" (UID: "678f61cd-297a-469e-b897-ec94ec29daf0"). InnerVolumeSpecName "kube-api-access-swqn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.195507 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config" (OuterVolumeSpecName: "config") pod "678f61cd-297a-469e-b897-ec94ec29daf0" (UID: "678f61cd-297a-469e-b897-ec94ec29daf0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.207835 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "678f61cd-297a-469e-b897-ec94ec29daf0" (UID: "678f61cd-297a-469e-b897-ec94ec29daf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.273276 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.273317 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678f61cd-297a-469e-b897-ec94ec29daf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.273331 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swqn4\" (UniqueName: \"kubernetes.io/projected/678f61cd-297a-469e-b897-ec94ec29daf0-kube-api-access-swqn4\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.781707 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lhdlc" event={"ID":"678f61cd-297a-469e-b897-ec94ec29daf0","Type":"ContainerDied","Data":"4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365"} Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.781750 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a75ef84ce497eb82c7afc09346fd81ef86762289209be81326922510d657365" Oct 06 10:11:47 crc kubenswrapper[4989]: I1006 10:11:47.781820 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lhdlc" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.032070 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:11:48 crc kubenswrapper[4989]: E1006 10:11:48.033139 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678f61cd-297a-469e-b897-ec94ec29daf0" containerName="neutron-db-sync" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.033155 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="678f61cd-297a-469e-b897-ec94ec29daf0" containerName="neutron-db-sync" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.034576 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="678f61cd-297a-469e-b897-ec94ec29daf0" containerName="neutron-db-sync" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.037442 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.053140 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.065480 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7954959975-7hbqd"] Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.067357 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.072601 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.072955 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qbb8h" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.074009 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.076749 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7954959975-7hbqd"] Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.085841 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086019 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dhc4\" (UniqueName: \"kubernetes.io/projected/0d12ecc3-3794-407e-a6e7-38198dc82213-kube-api-access-7dhc4\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086247 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086353 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gpmx\" (UniqueName: \"kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-combined-ca-bundle\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086515 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086632 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.086751 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-httpd-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187334 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187376 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dhc4\" (UniqueName: \"kubernetes.io/projected/0d12ecc3-3794-407e-a6e7-38198dc82213-kube-api-access-7dhc4\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187433 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187451 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gpmx\" (UniqueName: \"kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187472 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-combined-ca-bundle\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187494 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187525 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187542 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-httpd-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.187576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.189638 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.189637 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.190295 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.190509 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.194685 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.197929 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-httpd-config\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.208291 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dhc4\" (UniqueName: \"kubernetes.io/projected/0d12ecc3-3794-407e-a6e7-38198dc82213-kube-api-access-7dhc4\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.208930 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d12ecc3-3794-407e-a6e7-38198dc82213-combined-ca-bundle\") pod \"neutron-7954959975-7hbqd\" (UID: \"0d12ecc3-3794-407e-a6e7-38198dc82213\") " pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.211225 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gpmx\" (UniqueName: \"kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx\") pod \"dnsmasq-dns-695d5c49-4crkj\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.373236 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.387618 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:48 crc kubenswrapper[4989]: I1006 10:11:48.958651 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7954959975-7hbqd"] Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.095857 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:11:49 crc kubenswrapper[4989]: W1006 10:11:49.100828 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7663326d_5ee1_4702_9410_58e126e5bbbb.slice/crio-8bd96bf7cfcf75d847c5091cb17af3c164409f7f79f3c64515612e8c3c636ab8 WatchSource:0}: Error finding container 8bd96bf7cfcf75d847c5091cb17af3c164409f7f79f3c64515612e8c3c636ab8: Status 404 returned error can't find the container with id 8bd96bf7cfcf75d847c5091cb17af3c164409f7f79f3c64515612e8c3c636ab8 Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.615696 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.617078 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.670205 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.798266 4989 generic.go:334] "Generic (PLEG): container finished" podID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerID="ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4" exitCode=0 Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.798327 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-695d5c49-4crkj" event={"ID":"7663326d-5ee1-4702-9410-58e126e5bbbb","Type":"ContainerDied","Data":"ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4"} Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.798353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-695d5c49-4crkj" event={"ID":"7663326d-5ee1-4702-9410-58e126e5bbbb","Type":"ContainerStarted","Data":"8bd96bf7cfcf75d847c5091cb17af3c164409f7f79f3c64515612e8c3c636ab8"} Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.800877 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7954959975-7hbqd" event={"ID":"0d12ecc3-3794-407e-a6e7-38198dc82213","Type":"ContainerStarted","Data":"1ace08632d905c9b43610b69a5afeefb554772871f61331752f996c1786e357e"} Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.800927 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7954959975-7hbqd" event={"ID":"0d12ecc3-3794-407e-a6e7-38198dc82213","Type":"ContainerStarted","Data":"af888381618a0a621bbb4cff0d655280f337dcda502f7300faa3aa2b4d414061"} Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.800946 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7954959975-7hbqd" event={"ID":"0d12ecc3-3794-407e-a6e7-38198dc82213","Type":"ContainerStarted","Data":"0282f5fc49938d321ae0877e1fafc6b575adcaa678301a27743f22b312819e62"} Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.801274 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.853628 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7954959975-7hbqd" podStartSLOduration=1.8536047070000001 podStartE2EDuration="1.853604707s" podCreationTimestamp="2025-10-06 10:11:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:49.853291618 +0000 UTC m=+5560.643317198" watchObservedRunningTime="2025-10-06 10:11:49.853604707 +0000 UTC m=+5560.643630307" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.872393 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:49 crc kubenswrapper[4989]: I1006 10:11:49.953706 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:50 crc kubenswrapper[4989]: I1006 10:11:50.810428 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-695d5c49-4crkj" event={"ID":"7663326d-5ee1-4702-9410-58e126e5bbbb","Type":"ContainerStarted","Data":"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce"} Oct 06 10:11:50 crc kubenswrapper[4989]: I1006 10:11:50.834075 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-695d5c49-4crkj" podStartSLOduration=3.8340568790000003 podStartE2EDuration="3.834056879s" podCreationTimestamp="2025-10-06 10:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:11:50.828087737 +0000 UTC m=+5561.618113327" watchObservedRunningTime="2025-10-06 10:11:50.834056879 +0000 UTC m=+5561.624082459" Oct 06 10:11:51 crc kubenswrapper[4989]: I1006 10:11:51.817962 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gtx8c" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="registry-server" containerID="cri-o://6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e" gracePeriod=2 Oct 06 10:11:51 crc kubenswrapper[4989]: I1006 10:11:51.818418 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.275882 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.358803 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtrdp\" (UniqueName: \"kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp\") pod \"8878b926-e5c2-4c44-8858-615fa2aeb92f\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.358895 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content\") pod \"8878b926-e5c2-4c44-8858-615fa2aeb92f\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.359093 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities\") pod \"8878b926-e5c2-4c44-8858-615fa2aeb92f\" (UID: \"8878b926-e5c2-4c44-8858-615fa2aeb92f\") " Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.360288 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities" (OuterVolumeSpecName: "utilities") pod "8878b926-e5c2-4c44-8858-615fa2aeb92f" (UID: "8878b926-e5c2-4c44-8858-615fa2aeb92f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.364417 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp" (OuterVolumeSpecName: "kube-api-access-rtrdp") pod "8878b926-e5c2-4c44-8858-615fa2aeb92f" (UID: "8878b926-e5c2-4c44-8858-615fa2aeb92f"). InnerVolumeSpecName "kube-api-access-rtrdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.460755 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.460786 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtrdp\" (UniqueName: \"kubernetes.io/projected/8878b926-e5c2-4c44-8858-615fa2aeb92f-kube-api-access-rtrdp\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.826872 4989 generic.go:334] "Generic (PLEG): container finished" podID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerID="6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e" exitCode=0 Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.826962 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtx8c" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.826960 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerDied","Data":"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e"} Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.827020 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtx8c" event={"ID":"8878b926-e5c2-4c44-8858-615fa2aeb92f","Type":"ContainerDied","Data":"dfe088b480547bb8e53a0e407a68d73f0fb61d454f1e8892db0e36b82dd50164"} Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.827052 4989 scope.go:117] "RemoveContainer" containerID="6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.851221 4989 scope.go:117] "RemoveContainer" containerID="5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.875139 4989 scope.go:117] "RemoveContainer" containerID="2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.922765 4989 scope.go:117] "RemoveContainer" containerID="6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e" Oct 06 10:11:52 crc kubenswrapper[4989]: E1006 10:11:52.923183 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e\": container with ID starting with 6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e not found: ID does not exist" containerID="6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.923260 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e"} err="failed to get container status \"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e\": rpc error: code = NotFound desc = could not find container \"6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e\": container with ID starting with 6cc412095d8ed48bfc73bf36c39a6966842b0ec1b900eee2058726ede1c8787e not found: ID does not exist" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.923295 4989 scope.go:117] "RemoveContainer" containerID="5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222" Oct 06 10:11:52 crc kubenswrapper[4989]: E1006 10:11:52.923791 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222\": container with ID starting with 5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222 not found: ID does not exist" containerID="5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.923933 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222"} err="failed to get container status \"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222\": rpc error: code = NotFound desc = could not find container \"5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222\": container with ID starting with 5d576c354f6ddb5a3d8a0e53f23720f26c165c8201ea4cd65cba90976decd222 not found: ID does not exist" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.924127 4989 scope.go:117] "RemoveContainer" containerID="2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8" Oct 06 10:11:52 crc kubenswrapper[4989]: E1006 10:11:52.924819 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8\": container with ID starting with 2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8 not found: ID does not exist" containerID="2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8" Oct 06 10:11:52 crc kubenswrapper[4989]: I1006 10:11:52.924855 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8"} err="failed to get container status \"2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8\": rpc error: code = NotFound desc = could not find container \"2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8\": container with ID starting with 2d57891239ad8fd386bf44896f0b33b43da0a3811c51f3efa8925426d995dfe8 not found: ID does not exist" Oct 06 10:11:53 crc kubenswrapper[4989]: I1006 10:11:53.605531 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8878b926-e5c2-4c44-8858-615fa2aeb92f" (UID: "8878b926-e5c2-4c44-8858-615fa2aeb92f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:11:53 crc kubenswrapper[4989]: I1006 10:11:53.682271 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8878b926-e5c2-4c44-8858-615fa2aeb92f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:53 crc kubenswrapper[4989]: I1006 10:11:53.754344 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:53 crc kubenswrapper[4989]: I1006 10:11:53.762147 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gtx8c"] Oct 06 10:11:53 crc kubenswrapper[4989]: I1006 10:11:53.944693 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" path="/var/lib/kubelet/pods/8878b926-e5c2-4c44-8858-615fa2aeb92f/volumes" Oct 06 10:11:54 crc kubenswrapper[4989]: I1006 10:11:54.937409 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:11:54 crc kubenswrapper[4989]: E1006 10:11:54.937723 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.375945 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.436037 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.436340 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="dnsmasq-dns" containerID="cri-o://5e7d035e7a8d78246173cd60a429bf923ae69673a36dbad70fc327a5bb20ffaa" gracePeriod=10 Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.906622 4989 generic.go:334] "Generic (PLEG): container finished" podID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerID="5e7d035e7a8d78246173cd60a429bf923ae69673a36dbad70fc327a5bb20ffaa" exitCode=0 Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.906721 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" event={"ID":"7013d8e4-6361-41bf-8b32-82d3a2b7e557","Type":"ContainerDied","Data":"5e7d035e7a8d78246173cd60a429bf923ae69673a36dbad70fc327a5bb20ffaa"} Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.907079 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" event={"ID":"7013d8e4-6361-41bf-8b32-82d3a2b7e557","Type":"ContainerDied","Data":"0df2b0e870e3eaf7465ed8524c0ccbf11ae3f59249c747de404ed4299eda7836"} Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.907096 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0df2b0e870e3eaf7465ed8524c0ccbf11ae3f59249c747de404ed4299eda7836" Oct 06 10:11:58 crc kubenswrapper[4989]: I1006 10:11:58.910463 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.076604 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc\") pod \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.076675 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss9vr\" (UniqueName: \"kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr\") pod \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.076711 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config\") pod \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.076770 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb\") pod \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.076824 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb\") pod \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\" (UID: \"7013d8e4-6361-41bf-8b32-82d3a2b7e557\") " Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.084309 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr" (OuterVolumeSpecName: "kube-api-access-ss9vr") pod "7013d8e4-6361-41bf-8b32-82d3a2b7e557" (UID: "7013d8e4-6361-41bf-8b32-82d3a2b7e557"). InnerVolumeSpecName "kube-api-access-ss9vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.119623 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7013d8e4-6361-41bf-8b32-82d3a2b7e557" (UID: "7013d8e4-6361-41bf-8b32-82d3a2b7e557"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.120511 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config" (OuterVolumeSpecName: "config") pod "7013d8e4-6361-41bf-8b32-82d3a2b7e557" (UID: "7013d8e4-6361-41bf-8b32-82d3a2b7e557"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.136694 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7013d8e4-6361-41bf-8b32-82d3a2b7e557" (UID: "7013d8e4-6361-41bf-8b32-82d3a2b7e557"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.143904 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7013d8e4-6361-41bf-8b32-82d3a2b7e557" (UID: "7013d8e4-6361-41bf-8b32-82d3a2b7e557"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.179343 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.179409 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss9vr\" (UniqueName: \"kubernetes.io/projected/7013d8e4-6361-41bf-8b32-82d3a2b7e557-kube-api-access-ss9vr\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.179429 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.179443 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.179456 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7013d8e4-6361-41bf-8b32-82d3a2b7e557-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.913819 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774d46d7f7-bmc7j" Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.964638 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:11:59 crc kubenswrapper[4989]: I1006 10:11:59.972536 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-774d46d7f7-bmc7j"] Oct 06 10:12:01 crc kubenswrapper[4989]: I1006 10:12:01.951529 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" path="/var/lib/kubelet/pods/7013d8e4-6361-41bf-8b32-82d3a2b7e557/volumes" Oct 06 10:12:09 crc kubenswrapper[4989]: I1006 10:12:09.945841 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:12:09 crc kubenswrapper[4989]: E1006 10:12:09.947003 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:12:18 crc kubenswrapper[4989]: I1006 10:12:18.399944 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7954959975-7hbqd" Oct 06 10:12:20 crc kubenswrapper[4989]: E1006 10:12:20.858187 4989 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.23:55198->38.129.56.23:40321: write tcp 38.129.56.23:55198->38.129.56.23:40321: write: broken pipe Oct 06 10:12:22 crc kubenswrapper[4989]: I1006 10:12:22.936398 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:12:22 crc kubenswrapper[4989]: E1006 10:12:22.936863 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.427920 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-d2dmk"] Oct 06 10:12:26 crc kubenswrapper[4989]: E1006 10:12:26.428833 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="extract-utilities" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.428849 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="extract-utilities" Oct 06 10:12:26 crc kubenswrapper[4989]: E1006 10:12:26.428869 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="registry-server" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.428877 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="registry-server" Oct 06 10:12:26 crc kubenswrapper[4989]: E1006 10:12:26.428894 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="extract-content" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.428903 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="extract-content" Oct 06 10:12:26 crc kubenswrapper[4989]: E1006 10:12:26.428917 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="dnsmasq-dns" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.428924 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="dnsmasq-dns" Oct 06 10:12:26 crc kubenswrapper[4989]: E1006 10:12:26.428964 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="init" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.428971 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="init" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.429163 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8878b926-e5c2-4c44-8858-615fa2aeb92f" containerName="registry-server" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.429181 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7013d8e4-6361-41bf-8b32-82d3a2b7e557" containerName="dnsmasq-dns" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.429861 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.437951 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d2dmk"] Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.595710 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2dfm\" (UniqueName: \"kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm\") pod \"glance-db-create-d2dmk\" (UID: \"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8\") " pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.697589 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2dfm\" (UniqueName: \"kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm\") pod \"glance-db-create-d2dmk\" (UID: \"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8\") " pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.738550 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2dfm\" (UniqueName: \"kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm\") pod \"glance-db-create-d2dmk\" (UID: \"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8\") " pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:26 crc kubenswrapper[4989]: I1006 10:12:26.752674 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:27 crc kubenswrapper[4989]: I1006 10:12:27.197032 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d2dmk"] Oct 06 10:12:28 crc kubenswrapper[4989]: I1006 10:12:28.184336 4989 generic.go:334] "Generic (PLEG): container finished" podID="90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" containerID="380b5ba3224d49a023b83048b3c208b2163059125638d24a847532795a28b5b6" exitCode=0 Oct 06 10:12:28 crc kubenswrapper[4989]: I1006 10:12:28.184407 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d2dmk" event={"ID":"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8","Type":"ContainerDied","Data":"380b5ba3224d49a023b83048b3c208b2163059125638d24a847532795a28b5b6"} Oct 06 10:12:28 crc kubenswrapper[4989]: I1006 10:12:28.184887 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d2dmk" event={"ID":"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8","Type":"ContainerStarted","Data":"2c1ae319ffd412f52328690ff6f67f6a197aa0053483ea8ce9e14b46b3ae9551"} Oct 06 10:12:29 crc kubenswrapper[4989]: I1006 10:12:29.575468 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:29 crc kubenswrapper[4989]: I1006 10:12:29.748445 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2dfm\" (UniqueName: \"kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm\") pod \"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8\" (UID: \"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8\") " Oct 06 10:12:29 crc kubenswrapper[4989]: I1006 10:12:29.758324 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm" (OuterVolumeSpecName: "kube-api-access-v2dfm") pod "90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" (UID: "90ca868d-9f29-4ec0-8ca1-dbb6403bebb8"). InnerVolumeSpecName "kube-api-access-v2dfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:29 crc kubenswrapper[4989]: I1006 10:12:29.850829 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2dfm\" (UniqueName: \"kubernetes.io/projected/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8-kube-api-access-v2dfm\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:30 crc kubenswrapper[4989]: I1006 10:12:30.209978 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d2dmk" event={"ID":"90ca868d-9f29-4ec0-8ca1-dbb6403bebb8","Type":"ContainerDied","Data":"2c1ae319ffd412f52328690ff6f67f6a197aa0053483ea8ce9e14b46b3ae9551"} Oct 06 10:12:30 crc kubenswrapper[4989]: I1006 10:12:30.210036 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c1ae319ffd412f52328690ff6f67f6a197aa0053483ea8ce9e14b46b3ae9551" Oct 06 10:12:30 crc kubenswrapper[4989]: I1006 10:12:30.210132 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d2dmk" Oct 06 10:12:34 crc kubenswrapper[4989]: I1006 10:12:34.935803 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:12:35 crc kubenswrapper[4989]: I1006 10:12:35.260823 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0"} Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.507926 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fe7d-account-create-qsrmz"] Oct 06 10:12:36 crc kubenswrapper[4989]: E1006 10:12:36.509121 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" containerName="mariadb-database-create" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.509134 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" containerName="mariadb-database-create" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.509330 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" containerName="mariadb-database-create" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.510205 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.512683 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.523815 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fe7d-account-create-qsrmz"] Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.679563 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm7mt\" (UniqueName: \"kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt\") pod \"glance-fe7d-account-create-qsrmz\" (UID: \"21f0d9a4-084a-42e0-86ae-4dba3793472b\") " pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.781320 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm7mt\" (UniqueName: \"kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt\") pod \"glance-fe7d-account-create-qsrmz\" (UID: \"21f0d9a4-084a-42e0-86ae-4dba3793472b\") " pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.797734 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm7mt\" (UniqueName: \"kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt\") pod \"glance-fe7d-account-create-qsrmz\" (UID: \"21f0d9a4-084a-42e0-86ae-4dba3793472b\") " pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:36 crc kubenswrapper[4989]: I1006 10:12:36.836878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.050470 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.053004 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.082637 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.187904 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.187981 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mcch\" (UniqueName: \"kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.188302 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.291896 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.292210 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mcch\" (UniqueName: \"kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.292435 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.292630 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.292972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.298083 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fe7d-account-create-qsrmz"] Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.324469 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mcch\" (UniqueName: \"kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch\") pod \"certified-operators-snb7n\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.391515 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:37 crc kubenswrapper[4989]: I1006 10:12:37.655760 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:37 crc kubenswrapper[4989]: W1006 10:12:37.664806 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod018f7356_0c4a_4acf_a1b6_3e810151df3d.slice/crio-458c3037ff8fc13c432708bf0d91a74935c4b24d9547e002e9687679d2357f17 WatchSource:0}: Error finding container 458c3037ff8fc13c432708bf0d91a74935c4b24d9547e002e9687679d2357f17: Status 404 returned error can't find the container with id 458c3037ff8fc13c432708bf0d91a74935c4b24d9547e002e9687679d2357f17 Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.286590 4989 generic.go:334] "Generic (PLEG): container finished" podID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerID="35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e" exitCode=0 Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.286930 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerDied","Data":"35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e"} Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.286956 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerStarted","Data":"458c3037ff8fc13c432708bf0d91a74935c4b24d9547e002e9687679d2357f17"} Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.290672 4989 generic.go:334] "Generic (PLEG): container finished" podID="21f0d9a4-084a-42e0-86ae-4dba3793472b" containerID="d357906a4ae8b63392d01203589675721f42273e365cd5423e87aeb0670c0423" exitCode=0 Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.290714 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe7d-account-create-qsrmz" event={"ID":"21f0d9a4-084a-42e0-86ae-4dba3793472b","Type":"ContainerDied","Data":"d357906a4ae8b63392d01203589675721f42273e365cd5423e87aeb0670c0423"} Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.290740 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe7d-account-create-qsrmz" event={"ID":"21f0d9a4-084a-42e0-86ae-4dba3793472b","Type":"ContainerStarted","Data":"d4852b830e04385c784a6a0e18347514c477a2b4f6ba09bb3f597f0fac0068cf"} Oct 06 10:12:38 crc kubenswrapper[4989]: I1006 10:12:38.291515 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.300146 4989 generic.go:334] "Generic (PLEG): container finished" podID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerID="54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299" exitCode=0 Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.300202 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerDied","Data":"54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299"} Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.629597 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.737742 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm7mt\" (UniqueName: \"kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt\") pod \"21f0d9a4-084a-42e0-86ae-4dba3793472b\" (UID: \"21f0d9a4-084a-42e0-86ae-4dba3793472b\") " Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.746923 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt" (OuterVolumeSpecName: "kube-api-access-wm7mt") pod "21f0d9a4-084a-42e0-86ae-4dba3793472b" (UID: "21f0d9a4-084a-42e0-86ae-4dba3793472b"). InnerVolumeSpecName "kube-api-access-wm7mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:39 crc kubenswrapper[4989]: I1006 10:12:39.840641 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm7mt\" (UniqueName: \"kubernetes.io/projected/21f0d9a4-084a-42e0-86ae-4dba3793472b-kube-api-access-wm7mt\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:40 crc kubenswrapper[4989]: I1006 10:12:40.314844 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerStarted","Data":"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724"} Oct 06 10:12:40 crc kubenswrapper[4989]: I1006 10:12:40.316697 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe7d-account-create-qsrmz" event={"ID":"21f0d9a4-084a-42e0-86ae-4dba3793472b","Type":"ContainerDied","Data":"d4852b830e04385c784a6a0e18347514c477a2b4f6ba09bb3f597f0fac0068cf"} Oct 06 10:12:40 crc kubenswrapper[4989]: I1006 10:12:40.316732 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4852b830e04385c784a6a0e18347514c477a2b4f6ba09bb3f597f0fac0068cf" Oct 06 10:12:40 crc kubenswrapper[4989]: I1006 10:12:40.316776 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe7d-account-create-qsrmz" Oct 06 10:12:40 crc kubenswrapper[4989]: I1006 10:12:40.338392 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-snb7n" podStartSLOduration=1.9294512369999999 podStartE2EDuration="3.338371401s" podCreationTimestamp="2025-10-06 10:12:37 +0000 UTC" firstStartedPulling="2025-10-06 10:12:38.291229027 +0000 UTC m=+5609.081254607" lastFinishedPulling="2025-10-06 10:12:39.700149181 +0000 UTC m=+5610.490174771" observedRunningTime="2025-10-06 10:12:40.331124022 +0000 UTC m=+5611.121149602" watchObservedRunningTime="2025-10-06 10:12:40.338371401 +0000 UTC m=+5611.128396981" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.554476 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-rbnm2"] Oct 06 10:12:41 crc kubenswrapper[4989]: E1006 10:12:41.555036 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f0d9a4-084a-42e0-86ae-4dba3793472b" containerName="mariadb-account-create" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.555048 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f0d9a4-084a-42e0-86ae-4dba3793472b" containerName="mariadb-account-create" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.555207 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f0d9a4-084a-42e0-86ae-4dba3793472b" containerName="mariadb-account-create" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.555793 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.560436 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.562173 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lh4sv" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.569775 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rbnm2"] Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.669681 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.669793 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.669843 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.669885 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjdd5\" (UniqueName: \"kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.771521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.771799 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.771928 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjdd5\" (UniqueName: \"kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.772081 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.780387 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.781077 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.789184 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.799738 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjdd5\" (UniqueName: \"kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5\") pod \"glance-db-sync-rbnm2\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:41 crc kubenswrapper[4989]: I1006 10:12:41.898318 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:42 crc kubenswrapper[4989]: I1006 10:12:42.424216 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rbnm2"] Oct 06 10:12:42 crc kubenswrapper[4989]: W1006 10:12:42.429985 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ee627a_7a93_4ffc_b36f_f8c4b0a06839.slice/crio-7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63 WatchSource:0}: Error finding container 7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63: Status 404 returned error can't find the container with id 7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63 Oct 06 10:12:43 crc kubenswrapper[4989]: I1006 10:12:43.350095 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbnm2" event={"ID":"32ee627a-7a93-4ffc-b36f-f8c4b0a06839","Type":"ContainerStarted","Data":"9ac8bc981a972f516583e07399e57dfabc4e453a08b68e32a699ab23d85c1986"} Oct 06 10:12:43 crc kubenswrapper[4989]: I1006 10:12:43.350433 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbnm2" event={"ID":"32ee627a-7a93-4ffc-b36f-f8c4b0a06839","Type":"ContainerStarted","Data":"7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63"} Oct 06 10:12:43 crc kubenswrapper[4989]: I1006 10:12:43.371080 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-rbnm2" podStartSLOduration=2.371058474 podStartE2EDuration="2.371058474s" podCreationTimestamp="2025-10-06 10:12:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:43.366614526 +0000 UTC m=+5614.156640116" watchObservedRunningTime="2025-10-06 10:12:43.371058474 +0000 UTC m=+5614.161084054" Oct 06 10:12:47 crc kubenswrapper[4989]: I1006 10:12:47.392224 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:47 crc kubenswrapper[4989]: I1006 10:12:47.393724 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:47 crc kubenswrapper[4989]: I1006 10:12:47.395586 4989 generic.go:334] "Generic (PLEG): container finished" podID="32ee627a-7a93-4ffc-b36f-f8c4b0a06839" containerID="9ac8bc981a972f516583e07399e57dfabc4e453a08b68e32a699ab23d85c1986" exitCode=0 Oct 06 10:12:47 crc kubenswrapper[4989]: I1006 10:12:47.395624 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbnm2" event={"ID":"32ee627a-7a93-4ffc-b36f-f8c4b0a06839","Type":"ContainerDied","Data":"9ac8bc981a972f516583e07399e57dfabc4e453a08b68e32a699ab23d85c1986"} Oct 06 10:12:47 crc kubenswrapper[4989]: I1006 10:12:47.452807 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.458851 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.527447 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.897811 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.996899 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data\") pod \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.996998 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle\") pod \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.997069 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjdd5\" (UniqueName: \"kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5\") pod \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " Oct 06 10:12:48 crc kubenswrapper[4989]: I1006 10:12:48.997155 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data\") pod \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\" (UID: \"32ee627a-7a93-4ffc-b36f-f8c4b0a06839\") " Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.002076 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5" (OuterVolumeSpecName: "kube-api-access-cjdd5") pod "32ee627a-7a93-4ffc-b36f-f8c4b0a06839" (UID: "32ee627a-7a93-4ffc-b36f-f8c4b0a06839"). InnerVolumeSpecName "kube-api-access-cjdd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.004357 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32ee627a-7a93-4ffc-b36f-f8c4b0a06839" (UID: "32ee627a-7a93-4ffc-b36f-f8c4b0a06839"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.024270 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32ee627a-7a93-4ffc-b36f-f8c4b0a06839" (UID: "32ee627a-7a93-4ffc-b36f-f8c4b0a06839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.043366 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data" (OuterVolumeSpecName: "config-data") pod "32ee627a-7a93-4ffc-b36f-f8c4b0a06839" (UID: "32ee627a-7a93-4ffc-b36f-f8c4b0a06839"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.098997 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.099031 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.099043 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.099055 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjdd5\" (UniqueName: \"kubernetes.io/projected/32ee627a-7a93-4ffc-b36f-f8c4b0a06839-kube-api-access-cjdd5\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.410770 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rbnm2" event={"ID":"32ee627a-7a93-4ffc-b36f-f8c4b0a06839","Type":"ContainerDied","Data":"7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63"} Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.410822 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7272071f43e0925c712e19058459a20e8c5f6fa0f679dcfc803d1ef688ed9d63" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.410852 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rbnm2" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.674180 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:12:49 crc kubenswrapper[4989]: E1006 10:12:49.675114 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ee627a-7a93-4ffc-b36f-f8c4b0a06839" containerName="glance-db-sync" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.675134 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ee627a-7a93-4ffc-b36f-f8c4b0a06839" containerName="glance-db-sync" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.675362 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ee627a-7a93-4ffc-b36f-f8c4b0a06839" containerName="glance-db-sync" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.676475 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.685826 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.689410 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.699128 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.699185 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.699319 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lh4sv" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.699421 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.704063 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.718439 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.812953 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813006 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813035 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813062 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813082 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813115 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813141 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8r84\" (UniqueName: \"kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813196 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813240 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813261 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv8tt\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.813282 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.839836 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.844815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.847894 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.854309 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915441 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915609 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915639 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915842 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8r84\" (UniqueName: \"kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.915887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916123 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916224 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv8tt\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916283 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkj98\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916335 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916353 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916412 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916464 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916534 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916574 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916619 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916672 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916716 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.916758 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.917549 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.918155 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.918294 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.918583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.918882 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.920515 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.920550 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.920746 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.928691 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.932359 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.934194 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv8tt\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt\") pod \"glance-default-external-api-0\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:49 crc kubenswrapper[4989]: I1006 10:12:49.934341 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8r84\" (UniqueName: \"kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84\") pod \"dnsmasq-dns-78bb75bfd7-t92lf\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.005233 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018569 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018735 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkj98\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018759 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018796 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018831 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018856 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.018891 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.023331 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.023617 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.024241 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.024919 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.025559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.027912 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.028698 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.042219 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkj98\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98\") pod \"glance-default-internal-api-0\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.166980 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.335195 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.499788 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-snb7n" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="registry-server" containerID="cri-o://bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724" gracePeriod=2 Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.500402 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" event={"ID":"21448dfb-51d7-490e-bde9-c6195b66014f","Type":"ContainerStarted","Data":"f41135b67be14485b8a7150d383f6e7a5754ad079ef272370356f1011aa60c53"} Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.810367 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:50 crc kubenswrapper[4989]: W1006 10:12:50.823762 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaac75370_742f_4751_b836_3e5ba304bac7.slice/crio-ae3e09690cc9c389b7fb7acb443674598e34fb14567c6a4325100f9eb0053f9e WatchSource:0}: Error finding container ae3e09690cc9c389b7fb7acb443674598e34fb14567c6a4325100f9eb0053f9e: Status 404 returned error can't find the container with id ae3e09690cc9c389b7fb7acb443674598e34fb14567c6a4325100f9eb0053f9e Oct 06 10:12:50 crc kubenswrapper[4989]: I1006 10:12:50.952531 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.069069 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.145412 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content\") pod \"018f7356-0c4a-4acf-a1b6-3e810151df3d\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.145628 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mcch\" (UniqueName: \"kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch\") pod \"018f7356-0c4a-4acf-a1b6-3e810151df3d\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.145791 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities\") pod \"018f7356-0c4a-4acf-a1b6-3e810151df3d\" (UID: \"018f7356-0c4a-4acf-a1b6-3e810151df3d\") " Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.146622 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities" (OuterVolumeSpecName: "utilities") pod "018f7356-0c4a-4acf-a1b6-3e810151df3d" (UID: "018f7356-0c4a-4acf-a1b6-3e810151df3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.147209 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.151829 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch" (OuterVolumeSpecName: "kube-api-access-6mcch") pod "018f7356-0c4a-4acf-a1b6-3e810151df3d" (UID: "018f7356-0c4a-4acf-a1b6-3e810151df3d"). InnerVolumeSpecName "kube-api-access-6mcch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.204515 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "018f7356-0c4a-4acf-a1b6-3e810151df3d" (UID: "018f7356-0c4a-4acf-a1b6-3e810151df3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.249260 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mcch\" (UniqueName: \"kubernetes.io/projected/018f7356-0c4a-4acf-a1b6-3e810151df3d-kube-api-access-6mcch\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.249292 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/018f7356-0c4a-4acf-a1b6-3e810151df3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.348620 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.520071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" event={"ID":"21448dfb-51d7-490e-bde9-c6195b66014f","Type":"ContainerDied","Data":"f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.520097 4989 generic.go:334] "Generic (PLEG): container finished" podID="21448dfb-51d7-490e-bde9-c6195b66014f" containerID="f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae" exitCode=0 Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.523181 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerStarted","Data":"56e232ca2d090b7e2573f32d94715eb4cec05e2376f56072cc0477041dc3b564"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.530035 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerStarted","Data":"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.530070 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerStarted","Data":"ae3e09690cc9c389b7fb7acb443674598e34fb14567c6a4325100f9eb0053f9e"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.540939 4989 generic.go:334] "Generic (PLEG): container finished" podID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerID="bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724" exitCode=0 Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.540983 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerDied","Data":"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.541012 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-snb7n" event={"ID":"018f7356-0c4a-4acf-a1b6-3e810151df3d","Type":"ContainerDied","Data":"458c3037ff8fc13c432708bf0d91a74935c4b24d9547e002e9687679d2357f17"} Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.541030 4989 scope.go:117] "RemoveContainer" containerID="bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.541187 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-snb7n" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.572932 4989 scope.go:117] "RemoveContainer" containerID="54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.578018 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.601978 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-snb7n"] Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.607929 4989 scope.go:117] "RemoveContainer" containerID="35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.635267 4989 scope.go:117] "RemoveContainer" containerID="bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724" Oct 06 10:12:51 crc kubenswrapper[4989]: E1006 10:12:51.636177 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724\": container with ID starting with bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724 not found: ID does not exist" containerID="bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.636221 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724"} err="failed to get container status \"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724\": rpc error: code = NotFound desc = could not find container \"bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724\": container with ID starting with bf749566b0dea6ad3cdcc6f53b54f0653a48e8e6d54fe45d169adb283ad45724 not found: ID does not exist" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.636246 4989 scope.go:117] "RemoveContainer" containerID="54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299" Oct 06 10:12:51 crc kubenswrapper[4989]: E1006 10:12:51.641058 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299\": container with ID starting with 54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299 not found: ID does not exist" containerID="54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.641118 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299"} err="failed to get container status \"54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299\": rpc error: code = NotFound desc = could not find container \"54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299\": container with ID starting with 54716c008d0cfe8b63e5a29903c2e54064a560142dc282c1ea8999c181694299 not found: ID does not exist" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.641154 4989 scope.go:117] "RemoveContainer" containerID="35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e" Oct 06 10:12:51 crc kubenswrapper[4989]: E1006 10:12:51.641825 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e\": container with ID starting with 35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e not found: ID does not exist" containerID="35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.641872 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e"} err="failed to get container status \"35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e\": rpc error: code = NotFound desc = could not find container \"35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e\": container with ID starting with 35a921a0d639c8a2e4cd4d51bd15217731ea707b8f2b327009ea1fd865e4811e not found: ID does not exist" Oct 06 10:12:51 crc kubenswrapper[4989]: I1006 10:12:51.953630 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" path="/var/lib/kubelet/pods/018f7356-0c4a-4acf-a1b6-3e810151df3d/volumes" Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.556135 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" event={"ID":"21448dfb-51d7-490e-bde9-c6195b66014f","Type":"ContainerStarted","Data":"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2"} Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.557375 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.561001 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerStarted","Data":"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450"} Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.561030 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerStarted","Data":"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203"} Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.564323 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerStarted","Data":"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630"} Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.564436 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-log" containerID="cri-o://1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" gracePeriod=30 Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.564494 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-httpd" containerID="cri-o://8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" gracePeriod=30 Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.608996 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.608638279 podStartE2EDuration="3.608638279s" podCreationTimestamp="2025-10-06 10:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:52.599034502 +0000 UTC m=+5623.389060092" watchObservedRunningTime="2025-10-06 10:12:52.608638279 +0000 UTC m=+5623.398663879" Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.612415 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" podStartSLOduration=3.612397837 podStartE2EDuration="3.612397837s" podCreationTimestamp="2025-10-06 10:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:52.580089087 +0000 UTC m=+5623.370114677" watchObservedRunningTime="2025-10-06 10:12:52.612397837 +0000 UTC m=+5623.402423427" Oct 06 10:12:52 crc kubenswrapper[4989]: I1006 10:12:52.632628 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.6326118689999998 podStartE2EDuration="3.632611869s" podCreationTimestamp="2025-10-06 10:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:52.624882346 +0000 UTC m=+5623.414907926" watchObservedRunningTime="2025-10-06 10:12:52.632611869 +0000 UTC m=+5623.422637449" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.281038 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382083 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv8tt\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382276 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382326 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382349 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382394 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382414 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382445 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph\") pod \"aac75370-742f-4751-b836-3e5ba304bac7\" (UID: \"aac75370-742f-4751-b836-3e5ba304bac7\") " Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.382442 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.383439 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs" (OuterVolumeSpecName: "logs") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.389011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt" (OuterVolumeSpecName: "kube-api-access-jv8tt") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "kube-api-access-jv8tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.389156 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph" (OuterVolumeSpecName: "ceph") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.401897 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts" (OuterVolumeSpecName: "scripts") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.422604 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.431492 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data" (OuterVolumeSpecName: "config-data") pod "aac75370-742f-4751-b836-3e5ba304bac7" (UID: "aac75370-742f-4751-b836-3e5ba304bac7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484834 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484882 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv8tt\" (UniqueName: \"kubernetes.io/projected/aac75370-742f-4751-b836-3e5ba304bac7-kube-api-access-jv8tt\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484901 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484922 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484940 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aac75370-742f-4751-b836-3e5ba304bac7-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484955 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.484970 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aac75370-742f-4751-b836-3e5ba304bac7-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574225 4989 generic.go:334] "Generic (PLEG): container finished" podID="aac75370-742f-4751-b836-3e5ba304bac7" containerID="8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" exitCode=0 Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574256 4989 generic.go:334] "Generic (PLEG): container finished" podID="aac75370-742f-4751-b836-3e5ba304bac7" containerID="1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" exitCode=143 Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574309 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574316 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerDied","Data":"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630"} Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574377 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerDied","Data":"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29"} Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574390 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aac75370-742f-4751-b836-3e5ba304bac7","Type":"ContainerDied","Data":"ae3e09690cc9c389b7fb7acb443674598e34fb14567c6a4325100f9eb0053f9e"} Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.574407 4989 scope.go:117] "RemoveContainer" containerID="8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.609270 4989 scope.go:117] "RemoveContainer" containerID="1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.611237 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.624642 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.633555 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.634193 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="extract-utilities" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.634256 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="extract-utilities" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.634320 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-httpd" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.634397 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-httpd" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.634463 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="extract-content" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.634529 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="extract-content" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.634600 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-log" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.634666 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-log" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.634735 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="registry-server" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.634799 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="registry-server" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.635000 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-httpd" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.635075 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="018f7356-0c4a-4acf-a1b6-3e810151df3d" containerName="registry-server" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.635133 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac75370-742f-4751-b836-3e5ba304bac7" containerName="glance-log" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.636059 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.638583 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.652639 4989 scope.go:117] "RemoveContainer" containerID="8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.653060 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630\": container with ID starting with 8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630 not found: ID does not exist" containerID="8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653120 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630"} err="failed to get container status \"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630\": rpc error: code = NotFound desc = could not find container \"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630\": container with ID starting with 8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630 not found: ID does not exist" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653159 4989 scope.go:117] "RemoveContainer" containerID="1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" Oct 06 10:12:53 crc kubenswrapper[4989]: E1006 10:12:53.653462 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29\": container with ID starting with 1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29 not found: ID does not exist" containerID="1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653494 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29"} err="failed to get container status \"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29\": rpc error: code = NotFound desc = could not find container \"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29\": container with ID starting with 1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29 not found: ID does not exist" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653514 4989 scope.go:117] "RemoveContainer" containerID="8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653901 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630"} err="failed to get container status \"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630\": rpc error: code = NotFound desc = could not find container \"8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630\": container with ID starting with 8b40538cb780424ca1a9b0df3192767833afc387a3320f1880745aa05063e630 not found: ID does not exist" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.653922 4989 scope.go:117] "RemoveContainer" containerID="1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.654145 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29"} err="failed to get container status \"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29\": rpc error: code = NotFound desc = could not find container \"1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29\": container with ID starting with 1cfc3cfabe96907f7f52746143b411c14bc4bedeb191f018e5b44493f5647d29 not found: ID does not exist" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.657396 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.688801 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689068 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689203 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689288 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689348 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.689471 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9vs\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.700063 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.790583 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.790976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791007 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791028 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791068 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791094 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791134 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9vs\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791784 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.791879 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.794513 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.794710 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.794888 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.795063 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.806100 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9vs\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs\") pod \"glance-default-external-api-0\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " pod="openstack/glance-default-external-api-0" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.949240 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aac75370-742f-4751-b836-3e5ba304bac7" path="/var/lib/kubelet/pods/aac75370-742f-4751-b836-3e5ba304bac7/volumes" Oct 06 10:12:53 crc kubenswrapper[4989]: I1006 10:12:53.963567 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:12:54 crc kubenswrapper[4989]: I1006 10:12:54.491180 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:12:54 crc kubenswrapper[4989]: W1006 10:12:54.497587 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b488a4_65f3_446c_9e0b_7c78fb8ef126.slice/crio-4d82424efeeddf6e34d471014208d6f1b2d69d3c5088399d880d08f1c8855e82 WatchSource:0}: Error finding container 4d82424efeeddf6e34d471014208d6f1b2d69d3c5088399d880d08f1c8855e82: Status 404 returned error can't find the container with id 4d82424efeeddf6e34d471014208d6f1b2d69d3c5088399d880d08f1c8855e82 Oct 06 10:12:54 crc kubenswrapper[4989]: I1006 10:12:54.586071 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerStarted","Data":"4d82424efeeddf6e34d471014208d6f1b2d69d3c5088399d880d08f1c8855e82"} Oct 06 10:12:54 crc kubenswrapper[4989]: I1006 10:12:54.588096 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-log" containerID="cri-o://30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" gracePeriod=30 Oct 06 10:12:54 crc kubenswrapper[4989]: I1006 10:12:54.588148 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-httpd" containerID="cri-o://9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" gracePeriod=30 Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.185079 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.218908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219002 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219053 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219097 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219196 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219232 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.219276 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkj98\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98\") pod \"5c26a716-b032-45b0-abe0-b41f0a83326f\" (UID: \"5c26a716-b032-45b0-abe0-b41f0a83326f\") " Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.223990 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.224358 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98" (OuterVolumeSpecName: "kube-api-access-zkj98") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "kube-api-access-zkj98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.232791 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs" (OuterVolumeSpecName: "logs") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.233102 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts" (OuterVolumeSpecName: "scripts") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.237731 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph" (OuterVolumeSpecName: "ceph") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.249576 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.270053 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data" (OuterVolumeSpecName: "config-data") pod "5c26a716-b032-45b0-abe0-b41f0a83326f" (UID: "5c26a716-b032-45b0-abe0-b41f0a83326f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326043 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326189 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326200 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326211 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c26a716-b032-45b0-abe0-b41f0a83326f-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326219 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c26a716-b032-45b0-abe0-b41f0a83326f-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326227 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.326235 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkj98\" (UniqueName: \"kubernetes.io/projected/5c26a716-b032-45b0-abe0-b41f0a83326f-kube-api-access-zkj98\") on node \"crc\" DevicePath \"\"" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.601098 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerStarted","Data":"77f84f997ad1f7e98a8c6a63330f4bbb258d682434bff7e7e48b82b66af03cf3"} Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.601153 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerStarted","Data":"60ae53ed77dea2379dd474804638c0824f7e05563aaabb85a9c29c98eb00fa88"} Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603508 4989 generic.go:334] "Generic (PLEG): container finished" podID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerID="9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" exitCode=0 Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603548 4989 generic.go:334] "Generic (PLEG): container finished" podID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerID="30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" exitCode=143 Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603571 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerDied","Data":"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450"} Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603588 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603609 4989 scope.go:117] "RemoveContainer" containerID="9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603595 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerDied","Data":"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203"} Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.603919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c26a716-b032-45b0-abe0-b41f0a83326f","Type":"ContainerDied","Data":"56e232ca2d090b7e2573f32d94715eb4cec05e2376f56072cc0477041dc3b564"} Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.628954 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.628939106 podStartE2EDuration="2.628939106s" podCreationTimestamp="2025-10-06 10:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:55.624254261 +0000 UTC m=+5626.414279841" watchObservedRunningTime="2025-10-06 10:12:55.628939106 +0000 UTC m=+5626.418964686" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.643935 4989 scope.go:117] "RemoveContainer" containerID="30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.648529 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.656039 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.668942 4989 scope.go:117] "RemoveContainer" containerID="9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" Oct 06 10:12:55 crc kubenswrapper[4989]: E1006 10:12:55.669519 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450\": container with ID starting with 9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450 not found: ID does not exist" containerID="9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.669554 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450"} err="failed to get container status \"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450\": rpc error: code = NotFound desc = could not find container \"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450\": container with ID starting with 9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450 not found: ID does not exist" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.669578 4989 scope.go:117] "RemoveContainer" containerID="30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" Oct 06 10:12:55 crc kubenswrapper[4989]: E1006 10:12:55.670132 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203\": container with ID starting with 30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203 not found: ID does not exist" containerID="30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.670160 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203"} err="failed to get container status \"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203\": rpc error: code = NotFound desc = could not find container \"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203\": container with ID starting with 30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203 not found: ID does not exist" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.670177 4989 scope.go:117] "RemoveContainer" containerID="9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.670843 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450"} err="failed to get container status \"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450\": rpc error: code = NotFound desc = could not find container \"9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450\": container with ID starting with 9f7dbb1e7cd1b5f34b83d3861f5d436869a5639af45deb02511f524b56dd4450 not found: ID does not exist" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.670864 4989 scope.go:117] "RemoveContainer" containerID="30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.671048 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203"} err="failed to get container status \"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203\": rpc error: code = NotFound desc = could not find container \"30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203\": container with ID starting with 30a279b406d9dff54432c08b2d22d7a651ca2fd79425808e0d7faacbed467203 not found: ID does not exist" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.679369 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:55 crc kubenswrapper[4989]: E1006 10:12:55.679748 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-httpd" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.679762 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-httpd" Oct 06 10:12:55 crc kubenswrapper[4989]: E1006 10:12:55.679789 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-log" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.679795 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-log" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.679941 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-httpd" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.679965 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" containerName="glance-log" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.680865 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.682774 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.687928 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.733602 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.733744 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.733815 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.733869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.734246 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.734290 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmhd5\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.734307 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835754 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835804 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmhd5\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835868 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835913 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.835966 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.836007 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.836335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.836600 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.840610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.841316 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.841723 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.841814 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.853400 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmhd5\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5\") pod \"glance-default-internal-api-0\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:12:55 crc kubenswrapper[4989]: I1006 10:12:55.949221 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c26a716-b032-45b0-abe0-b41f0a83326f" path="/var/lib/kubelet/pods/5c26a716-b032-45b0-abe0-b41f0a83326f/volumes" Oct 06 10:12:56 crc kubenswrapper[4989]: I1006 10:12:56.000237 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:12:56 crc kubenswrapper[4989]: I1006 10:12:56.501380 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:12:56 crc kubenswrapper[4989]: W1006 10:12:56.515210 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44aa1eb6_a9b6_422f_a91f_27ef20555668.slice/crio-ecd4ff97f82c465d72ac8dd147790a03e2a8b9dc8f6aa7e13e78ce1692045f8a WatchSource:0}: Error finding container ecd4ff97f82c465d72ac8dd147790a03e2a8b9dc8f6aa7e13e78ce1692045f8a: Status 404 returned error can't find the container with id ecd4ff97f82c465d72ac8dd147790a03e2a8b9dc8f6aa7e13e78ce1692045f8a Oct 06 10:12:56 crc kubenswrapper[4989]: I1006 10:12:56.612998 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerStarted","Data":"ecd4ff97f82c465d72ac8dd147790a03e2a8b9dc8f6aa7e13e78ce1692045f8a"} Oct 06 10:12:57 crc kubenswrapper[4989]: I1006 10:12:57.624414 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerStarted","Data":"eb83642e14061395a0ab65f70817c6ab2e76e892d431600bbb05ad2460bd2101"} Oct 06 10:12:57 crc kubenswrapper[4989]: I1006 10:12:57.624795 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerStarted","Data":"3aeba07768ba71f5efff33027feece42bf9054252e82470640a86584b7aa923f"} Oct 06 10:12:57 crc kubenswrapper[4989]: I1006 10:12:57.672002 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.671985033 podStartE2EDuration="2.671985033s" podCreationTimestamp="2025-10-06 10:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:12:57.650209287 +0000 UTC m=+5628.440234917" watchObservedRunningTime="2025-10-06 10:12:57.671985033 +0000 UTC m=+5628.462010623" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.006907 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.090360 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.090619 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-695d5c49-4crkj" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="dnsmasq-dns" containerID="cri-o://089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce" gracePeriod=10 Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.577252 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.666572 4989 generic.go:334] "Generic (PLEG): container finished" podID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerID="089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce" exitCode=0 Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.666618 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-695d5c49-4crkj" event={"ID":"7663326d-5ee1-4702-9410-58e126e5bbbb","Type":"ContainerDied","Data":"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce"} Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.666687 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-695d5c49-4crkj" event={"ID":"7663326d-5ee1-4702-9410-58e126e5bbbb","Type":"ContainerDied","Data":"8bd96bf7cfcf75d847c5091cb17af3c164409f7f79f3c64515612e8c3c636ab8"} Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.666705 4989 scope.go:117] "RemoveContainer" containerID="089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.666744 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-695d5c49-4crkj" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.685249 4989 scope.go:117] "RemoveContainer" containerID="ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.703126 4989 scope.go:117] "RemoveContainer" containerID="089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce" Oct 06 10:13:00 crc kubenswrapper[4989]: E1006 10:13:00.703810 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce\": container with ID starting with 089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce not found: ID does not exist" containerID="089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.703864 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce"} err="failed to get container status \"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce\": rpc error: code = NotFound desc = could not find container \"089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce\": container with ID starting with 089777922e36d0d5e330652919ec5fcaf6689b6b8d5f2fc502e665d131b766ce not found: ID does not exist" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.703900 4989 scope.go:117] "RemoveContainer" containerID="ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4" Oct 06 10:13:00 crc kubenswrapper[4989]: E1006 10:13:00.704445 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4\": container with ID starting with ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4 not found: ID does not exist" containerID="ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.704518 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4"} err="failed to get container status \"ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4\": rpc error: code = NotFound desc = could not find container \"ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4\": container with ID starting with ad654bb517af887f9fdaec0def047716da469c539f1efe068d9ab7d57978cab4 not found: ID does not exist" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.727050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc\") pod \"7663326d-5ee1-4702-9410-58e126e5bbbb\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.727115 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb\") pod \"7663326d-5ee1-4702-9410-58e126e5bbbb\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.728067 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb\") pod \"7663326d-5ee1-4702-9410-58e126e5bbbb\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.728150 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gpmx\" (UniqueName: \"kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx\") pod \"7663326d-5ee1-4702-9410-58e126e5bbbb\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.728197 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config\") pod \"7663326d-5ee1-4702-9410-58e126e5bbbb\" (UID: \"7663326d-5ee1-4702-9410-58e126e5bbbb\") " Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.732783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx" (OuterVolumeSpecName: "kube-api-access-5gpmx") pod "7663326d-5ee1-4702-9410-58e126e5bbbb" (UID: "7663326d-5ee1-4702-9410-58e126e5bbbb"). InnerVolumeSpecName "kube-api-access-5gpmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.766975 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7663326d-5ee1-4702-9410-58e126e5bbbb" (UID: "7663326d-5ee1-4702-9410-58e126e5bbbb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.774055 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7663326d-5ee1-4702-9410-58e126e5bbbb" (UID: "7663326d-5ee1-4702-9410-58e126e5bbbb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.798296 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config" (OuterVolumeSpecName: "config") pod "7663326d-5ee1-4702-9410-58e126e5bbbb" (UID: "7663326d-5ee1-4702-9410-58e126e5bbbb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.799959 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7663326d-5ee1-4702-9410-58e126e5bbbb" (UID: "7663326d-5ee1-4702-9410-58e126e5bbbb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.830356 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.830402 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.830424 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gpmx\" (UniqueName: \"kubernetes.io/projected/7663326d-5ee1-4702-9410-58e126e5bbbb-kube-api-access-5gpmx\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.830520 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:00 crc kubenswrapper[4989]: I1006 10:13:00.830538 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7663326d-5ee1-4702-9410-58e126e5bbbb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:01 crc kubenswrapper[4989]: I1006 10:13:01.023986 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:13:01 crc kubenswrapper[4989]: I1006 10:13:01.035883 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-695d5c49-4crkj"] Oct 06 10:13:01 crc kubenswrapper[4989]: I1006 10:13:01.956960 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" path="/var/lib/kubelet/pods/7663326d-5ee1-4702-9410-58e126e5bbbb/volumes" Oct 06 10:13:03 crc kubenswrapper[4989]: I1006 10:13:03.964553 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:13:03 crc kubenswrapper[4989]: I1006 10:13:03.964615 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:13:04 crc kubenswrapper[4989]: I1006 10:13:04.010784 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:13:04 crc kubenswrapper[4989]: I1006 10:13:04.025751 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:13:04 crc kubenswrapper[4989]: I1006 10:13:04.707117 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:13:04 crc kubenswrapper[4989]: I1006 10:13:04.707553 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.001443 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.001494 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.032140 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.063936 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.599485 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.635725 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.724559 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:06 crc kubenswrapper[4989]: I1006 10:13:06.724599 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:08 crc kubenswrapper[4989]: I1006 10:13:08.669698 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:08 crc kubenswrapper[4989]: I1006 10:13:08.724472 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.945431 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-scfdn"] Oct 06 10:13:14 crc kubenswrapper[4989]: E1006 10:13:14.946084 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="dnsmasq-dns" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.946097 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="dnsmasq-dns" Oct 06 10:13:14 crc kubenswrapper[4989]: E1006 10:13:14.946122 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="init" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.946128 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="init" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.946311 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7663326d-5ee1-4702-9410-58e126e5bbbb" containerName="dnsmasq-dns" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.946943 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-scfdn" Oct 06 10:13:14 crc kubenswrapper[4989]: I1006 10:13:14.953337 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-scfdn"] Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.109234 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrc6q\" (UniqueName: \"kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q\") pod \"placement-db-create-scfdn\" (UID: \"36224a2c-ec62-4e04-8541-a09f8c2671ca\") " pod="openstack/placement-db-create-scfdn" Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.211418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrc6q\" (UniqueName: \"kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q\") pod \"placement-db-create-scfdn\" (UID: \"36224a2c-ec62-4e04-8541-a09f8c2671ca\") " pod="openstack/placement-db-create-scfdn" Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.238180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrc6q\" (UniqueName: \"kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q\") pod \"placement-db-create-scfdn\" (UID: \"36224a2c-ec62-4e04-8541-a09f8c2671ca\") " pod="openstack/placement-db-create-scfdn" Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.264238 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-scfdn" Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.740932 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-scfdn"] Oct 06 10:13:15 crc kubenswrapper[4989]: I1006 10:13:15.824257 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-scfdn" event={"ID":"36224a2c-ec62-4e04-8541-a09f8c2671ca","Type":"ContainerStarted","Data":"46987ad26002d92b70da32c658f39fcb85d7e256afbf91f970371f75116079fc"} Oct 06 10:13:16 crc kubenswrapper[4989]: I1006 10:13:16.839752 4989 generic.go:334] "Generic (PLEG): container finished" podID="36224a2c-ec62-4e04-8541-a09f8c2671ca" containerID="9de599edc0650199df1987772a836f935514ddefe755bcdea1642b15903bdc60" exitCode=0 Oct 06 10:13:16 crc kubenswrapper[4989]: I1006 10:13:16.839865 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-scfdn" event={"ID":"36224a2c-ec62-4e04-8541-a09f8c2671ca","Type":"ContainerDied","Data":"9de599edc0650199df1987772a836f935514ddefe755bcdea1642b15903bdc60"} Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.258351 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-scfdn" Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.369575 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrc6q\" (UniqueName: \"kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q\") pod \"36224a2c-ec62-4e04-8541-a09f8c2671ca\" (UID: \"36224a2c-ec62-4e04-8541-a09f8c2671ca\") " Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.376178 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q" (OuterVolumeSpecName: "kube-api-access-xrc6q") pod "36224a2c-ec62-4e04-8541-a09f8c2671ca" (UID: "36224a2c-ec62-4e04-8541-a09f8c2671ca"). InnerVolumeSpecName "kube-api-access-xrc6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.471200 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrc6q\" (UniqueName: \"kubernetes.io/projected/36224a2c-ec62-4e04-8541-a09f8c2671ca-kube-api-access-xrc6q\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.863898 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-scfdn" event={"ID":"36224a2c-ec62-4e04-8541-a09f8c2671ca","Type":"ContainerDied","Data":"46987ad26002d92b70da32c658f39fcb85d7e256afbf91f970371f75116079fc"} Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.863960 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46987ad26002d92b70da32c658f39fcb85d7e256afbf91f970371f75116079fc" Oct 06 10:13:18 crc kubenswrapper[4989]: I1006 10:13:18.863969 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-scfdn" Oct 06 10:13:20 crc kubenswrapper[4989]: I1006 10:13:20.790602 4989 scope.go:117] "RemoveContainer" containerID="8d8fb23139b64792e555c273cae30b610a82d5e23253fa4124f66c1e9d64794a" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.024112 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f4f3-account-create-vkrq9"] Oct 06 10:13:25 crc kubenswrapper[4989]: E1006 10:13:25.025123 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36224a2c-ec62-4e04-8541-a09f8c2671ca" containerName="mariadb-database-create" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.025140 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="36224a2c-ec62-4e04-8541-a09f8c2671ca" containerName="mariadb-database-create" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.025343 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="36224a2c-ec62-4e04-8541-a09f8c2671ca" containerName="mariadb-database-create" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.025903 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.028354 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.033719 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f4f3-account-create-vkrq9"] Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.097579 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vds6v\" (UniqueName: \"kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v\") pod \"placement-f4f3-account-create-vkrq9\" (UID: \"15025438-e8fe-460e-a65f-2776b76f0cbd\") " pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.199177 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vds6v\" (UniqueName: \"kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v\") pod \"placement-f4f3-account-create-vkrq9\" (UID: \"15025438-e8fe-460e-a65f-2776b76f0cbd\") " pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.223164 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vds6v\" (UniqueName: \"kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v\") pod \"placement-f4f3-account-create-vkrq9\" (UID: \"15025438-e8fe-460e-a65f-2776b76f0cbd\") " pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.352637 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:25 crc kubenswrapper[4989]: I1006 10:13:25.933862 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f4f3-account-create-vkrq9"] Oct 06 10:13:26 crc kubenswrapper[4989]: I1006 10:13:26.959223 4989 generic.go:334] "Generic (PLEG): container finished" podID="15025438-e8fe-460e-a65f-2776b76f0cbd" containerID="3e29d9cd39bbeec7ce629ec2784d177ff51d2c4665d5a2c5204512b3f850bab0" exitCode=0 Oct 06 10:13:26 crc kubenswrapper[4989]: I1006 10:13:26.959570 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f4f3-account-create-vkrq9" event={"ID":"15025438-e8fe-460e-a65f-2776b76f0cbd","Type":"ContainerDied","Data":"3e29d9cd39bbeec7ce629ec2784d177ff51d2c4665d5a2c5204512b3f850bab0"} Oct 06 10:13:26 crc kubenswrapper[4989]: I1006 10:13:26.959608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f4f3-account-create-vkrq9" event={"ID":"15025438-e8fe-460e-a65f-2776b76f0cbd","Type":"ContainerStarted","Data":"f970a6945d0eebdba383fa5d22a26e33238575663681eed45f113e5dc5a11e5a"} Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.303810 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.362805 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vds6v\" (UniqueName: \"kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v\") pod \"15025438-e8fe-460e-a65f-2776b76f0cbd\" (UID: \"15025438-e8fe-460e-a65f-2776b76f0cbd\") " Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.368347 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v" (OuterVolumeSpecName: "kube-api-access-vds6v") pod "15025438-e8fe-460e-a65f-2776b76f0cbd" (UID: "15025438-e8fe-460e-a65f-2776b76f0cbd"). InnerVolumeSpecName "kube-api-access-vds6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.464955 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vds6v\" (UniqueName: \"kubernetes.io/projected/15025438-e8fe-460e-a65f-2776b76f0cbd-kube-api-access-vds6v\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.984227 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f4f3-account-create-vkrq9" event={"ID":"15025438-e8fe-460e-a65f-2776b76f0cbd","Type":"ContainerDied","Data":"f970a6945d0eebdba383fa5d22a26e33238575663681eed45f113e5dc5a11e5a"} Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.984301 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f4f3-account-create-vkrq9" Oct 06 10:13:28 crc kubenswrapper[4989]: I1006 10:13:28.984312 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f970a6945d0eebdba383fa5d22a26e33238575663681eed45f113e5dc5a11e5a" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.249878 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:13:30 crc kubenswrapper[4989]: E1006 10:13:30.251278 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15025438-e8fe-460e-a65f-2776b76f0cbd" containerName="mariadb-account-create" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.251405 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="15025438-e8fe-460e-a65f-2776b76f0cbd" containerName="mariadb-account-create" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.251810 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="15025438-e8fe-460e-a65f-2776b76f0cbd" containerName="mariadb-account-create" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.253026 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.263174 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.300690 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.300949 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.301013 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.301037 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.301177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmnf4\" (UniqueName: \"kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.317124 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-2vkss"] Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.318176 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.320105 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.320275 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wtbg8" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.320388 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.324155 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vkss"] Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.402835 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.402893 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.402917 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.402939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk4dq\" (UniqueName: \"kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403034 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403065 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403091 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403118 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403135 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403177 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmnf4\" (UniqueName: \"kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403803 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.403839 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.404008 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.404210 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.424491 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmnf4\" (UniqueName: \"kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4\") pod \"dnsmasq-dns-545cdf9765-8hzmr\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.504944 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.504995 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.505017 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk4dq\" (UniqueName: \"kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.505060 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.505085 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.505578 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.508797 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.509374 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.512166 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.520708 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk4dq\" (UniqueName: \"kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq\") pod \"placement-db-sync-2vkss\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.570610 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.637071 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:30 crc kubenswrapper[4989]: I1006 10:13:30.800721 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:13:31 crc kubenswrapper[4989]: I1006 10:13:31.006148 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerStarted","Data":"63a6b31b4b99335a8445802e0285b1a876886407b9fa0b649033521e97818c7e"} Oct 06 10:13:31 crc kubenswrapper[4989]: I1006 10:13:31.006622 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerStarted","Data":"3dd0d9294036145712497635b9af000ac81ae8a1d79021e46f984db5fd38f192"} Oct 06 10:13:31 crc kubenswrapper[4989]: I1006 10:13:31.110556 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-2vkss"] Oct 06 10:13:31 crc kubenswrapper[4989]: W1006 10:13:31.119203 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode01ce706_6ac5_4f40_b472_c6ea75ff4d64.slice/crio-325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e WatchSource:0}: Error finding container 325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e: Status 404 returned error can't find the container with id 325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e Oct 06 10:13:32 crc kubenswrapper[4989]: I1006 10:13:32.020093 4989 generic.go:334] "Generic (PLEG): container finished" podID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerID="63a6b31b4b99335a8445802e0285b1a876886407b9fa0b649033521e97818c7e" exitCode=0 Oct 06 10:13:32 crc kubenswrapper[4989]: I1006 10:13:32.020220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerDied","Data":"63a6b31b4b99335a8445802e0285b1a876886407b9fa0b649033521e97818c7e"} Oct 06 10:13:32 crc kubenswrapper[4989]: I1006 10:13:32.058362 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vkss" event={"ID":"e01ce706-6ac5-4f40-b472-c6ea75ff4d64","Type":"ContainerStarted","Data":"426bf8118a2dbdc160b9328f182934de7b9703121cbbbea313ba7f037256bc98"} Oct 06 10:13:32 crc kubenswrapper[4989]: I1006 10:13:32.058413 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vkss" event={"ID":"e01ce706-6ac5-4f40-b472-c6ea75ff4d64","Type":"ContainerStarted","Data":"325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e"} Oct 06 10:13:32 crc kubenswrapper[4989]: I1006 10:13:32.090966 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-2vkss" podStartSLOduration=2.090938749 podStartE2EDuration="2.090938749s" podCreationTimestamp="2025-10-06 10:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:13:32.081162568 +0000 UTC m=+5662.871188168" watchObservedRunningTime="2025-10-06 10:13:32.090938749 +0000 UTC m=+5662.880964339" Oct 06 10:13:33 crc kubenswrapper[4989]: I1006 10:13:33.068754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerStarted","Data":"0a4bb2572669056cfaa0c535b365a11c343e72325a6e6741efd32ebcdaa35195"} Oct 06 10:13:33 crc kubenswrapper[4989]: I1006 10:13:33.069274 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:33 crc kubenswrapper[4989]: I1006 10:13:33.070791 4989 generic.go:334] "Generic (PLEG): container finished" podID="e01ce706-6ac5-4f40-b472-c6ea75ff4d64" containerID="426bf8118a2dbdc160b9328f182934de7b9703121cbbbea313ba7f037256bc98" exitCode=0 Oct 06 10:13:33 crc kubenswrapper[4989]: I1006 10:13:33.070829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vkss" event={"ID":"e01ce706-6ac5-4f40-b472-c6ea75ff4d64","Type":"ContainerDied","Data":"426bf8118a2dbdc160b9328f182934de7b9703121cbbbea313ba7f037256bc98"} Oct 06 10:13:33 crc kubenswrapper[4989]: I1006 10:13:33.090099 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" podStartSLOduration=3.090083199 podStartE2EDuration="3.090083199s" podCreationTimestamp="2025-10-06 10:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:13:33.087234777 +0000 UTC m=+5663.877260397" watchObservedRunningTime="2025-10-06 10:13:33.090083199 +0000 UTC m=+5663.880108779" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.406612 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.476565 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk4dq\" (UniqueName: \"kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq\") pod \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.476781 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data\") pod \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.476811 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle\") pod \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.476833 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs\") pod \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.476871 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts\") pod \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\" (UID: \"e01ce706-6ac5-4f40-b472-c6ea75ff4d64\") " Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.477378 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs" (OuterVolumeSpecName: "logs") pod "e01ce706-6ac5-4f40-b472-c6ea75ff4d64" (UID: "e01ce706-6ac5-4f40-b472-c6ea75ff4d64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.484015 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts" (OuterVolumeSpecName: "scripts") pod "e01ce706-6ac5-4f40-b472-c6ea75ff4d64" (UID: "e01ce706-6ac5-4f40-b472-c6ea75ff4d64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.484067 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq" (OuterVolumeSpecName: "kube-api-access-fk4dq") pod "e01ce706-6ac5-4f40-b472-c6ea75ff4d64" (UID: "e01ce706-6ac5-4f40-b472-c6ea75ff4d64"). InnerVolumeSpecName "kube-api-access-fk4dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.507793 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e01ce706-6ac5-4f40-b472-c6ea75ff4d64" (UID: "e01ce706-6ac5-4f40-b472-c6ea75ff4d64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.518126 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data" (OuterVolumeSpecName: "config-data") pod "e01ce706-6ac5-4f40-b472-c6ea75ff4d64" (UID: "e01ce706-6ac5-4f40-b472-c6ea75ff4d64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.579465 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.579494 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.579505 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.579513 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:34 crc kubenswrapper[4989]: I1006 10:13:34.579524 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk4dq\" (UniqueName: \"kubernetes.io/projected/e01ce706-6ac5-4f40-b472-c6ea75ff4d64-kube-api-access-fk4dq\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.096562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-2vkss" event={"ID":"e01ce706-6ac5-4f40-b472-c6ea75ff4d64","Type":"ContainerDied","Data":"325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e"} Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.096854 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="325b7f728e1e095d07cd60398357adf95eb2f9ca2d6010d9a417dc29df8b3b0e" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.096628 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-2vkss" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.166433 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-849964bbfd-whdvn"] Oct 06 10:13:35 crc kubenswrapper[4989]: E1006 10:13:35.166827 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e01ce706-6ac5-4f40-b472-c6ea75ff4d64" containerName="placement-db-sync" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.166847 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e01ce706-6ac5-4f40-b472-c6ea75ff4d64" containerName="placement-db-sync" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.167102 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e01ce706-6ac5-4f40-b472-c6ea75ff4d64" containerName="placement-db-sync" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.168159 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.170668 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.171068 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wtbg8" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.171459 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.183969 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-849964bbfd-whdvn"] Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.294292 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-scripts\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.294384 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee9b21b-081f-494d-b8de-7f4919a2d16c-logs\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.294425 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gl4j\" (UniqueName: \"kubernetes.io/projected/4ee9b21b-081f-494d-b8de-7f4919a2d16c-kube-api-access-5gl4j\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.294472 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-config-data\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.294505 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-combined-ca-bundle\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.396419 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-config-data\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.396469 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-combined-ca-bundle\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.396617 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-scripts\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.396703 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee9b21b-081f-494d-b8de-7f4919a2d16c-logs\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.396741 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gl4j\" (UniqueName: \"kubernetes.io/projected/4ee9b21b-081f-494d-b8de-7f4919a2d16c-kube-api-access-5gl4j\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.397735 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ee9b21b-081f-494d-b8de-7f4919a2d16c-logs\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.400332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-scripts\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.401830 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-config-data\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.412437 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee9b21b-081f-494d-b8de-7f4919a2d16c-combined-ca-bundle\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.416341 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gl4j\" (UniqueName: \"kubernetes.io/projected/4ee9b21b-081f-494d-b8de-7f4919a2d16c-kube-api-access-5gl4j\") pod \"placement-849964bbfd-whdvn\" (UID: \"4ee9b21b-081f-494d-b8de-7f4919a2d16c\") " pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.487062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:35 crc kubenswrapper[4989]: I1006 10:13:35.907762 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-849964bbfd-whdvn"] Oct 06 10:13:35 crc kubenswrapper[4989]: W1006 10:13:35.916172 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ee9b21b_081f_494d_b8de_7f4919a2d16c.slice/crio-146a6d4e0cb989d2b38ae9c664edd8ca55fc5b7730eb93c699a96e1b7c06297d WatchSource:0}: Error finding container 146a6d4e0cb989d2b38ae9c664edd8ca55fc5b7730eb93c699a96e1b7c06297d: Status 404 returned error can't find the container with id 146a6d4e0cb989d2b38ae9c664edd8ca55fc5b7730eb93c699a96e1b7c06297d Oct 06 10:13:36 crc kubenswrapper[4989]: I1006 10:13:36.110635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849964bbfd-whdvn" event={"ID":"4ee9b21b-081f-494d-b8de-7f4919a2d16c","Type":"ContainerStarted","Data":"bcd9591a854735c9b09f26e5c802953502ba3cd028105cbfa1cd28eda40ed732"} Oct 06 10:13:36 crc kubenswrapper[4989]: I1006 10:13:36.110954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849964bbfd-whdvn" event={"ID":"4ee9b21b-081f-494d-b8de-7f4919a2d16c","Type":"ContainerStarted","Data":"146a6d4e0cb989d2b38ae9c664edd8ca55fc5b7730eb93c699a96e1b7c06297d"} Oct 06 10:13:37 crc kubenswrapper[4989]: I1006 10:13:37.123064 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-849964bbfd-whdvn" event={"ID":"4ee9b21b-081f-494d-b8de-7f4919a2d16c","Type":"ContainerStarted","Data":"8dde1bc1697d520c8d5dbb21e38f927397023257ef6f9690828f8781f154ad1a"} Oct 06 10:13:37 crc kubenswrapper[4989]: I1006 10:13:37.124059 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:37 crc kubenswrapper[4989]: I1006 10:13:37.124161 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:13:37 crc kubenswrapper[4989]: I1006 10:13:37.153348 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-849964bbfd-whdvn" podStartSLOduration=2.153324835 podStartE2EDuration="2.153324835s" podCreationTimestamp="2025-10-06 10:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:13:37.149066892 +0000 UTC m=+5667.939092502" watchObservedRunningTime="2025-10-06 10:13:37.153324835 +0000 UTC m=+5667.943350445" Oct 06 10:13:40 crc kubenswrapper[4989]: I1006 10:13:40.572457 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:13:40 crc kubenswrapper[4989]: I1006 10:13:40.646443 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:13:40 crc kubenswrapper[4989]: I1006 10:13:40.646917 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="dnsmasq-dns" containerID="cri-o://47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2" gracePeriod=10 Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.125877 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.170988 4989 generic.go:334] "Generic (PLEG): container finished" podID="21448dfb-51d7-490e-bde9-c6195b66014f" containerID="47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2" exitCode=0 Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.171032 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.171029 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" event={"ID":"21448dfb-51d7-490e-bde9-c6195b66014f","Type":"ContainerDied","Data":"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2"} Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.171079 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78bb75bfd7-t92lf" event={"ID":"21448dfb-51d7-490e-bde9-c6195b66014f","Type":"ContainerDied","Data":"f41135b67be14485b8a7150d383f6e7a5754ad079ef272370356f1011aa60c53"} Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.171115 4989 scope.go:117] "RemoveContainer" containerID="47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.191829 4989 scope.go:117] "RemoveContainer" containerID="f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.207304 4989 scope.go:117] "RemoveContainer" containerID="47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2" Oct 06 10:13:41 crc kubenswrapper[4989]: E1006 10:13:41.207667 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2\": container with ID starting with 47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2 not found: ID does not exist" containerID="47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.207692 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2"} err="failed to get container status \"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2\": rpc error: code = NotFound desc = could not find container \"47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2\": container with ID starting with 47ad9d1e280aad65a9114a95099dd8b0bedcf52685a9fb91bff66bf73d5760a2 not found: ID does not exist" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.207710 4989 scope.go:117] "RemoveContainer" containerID="f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae" Oct 06 10:13:41 crc kubenswrapper[4989]: E1006 10:13:41.207989 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae\": container with ID starting with f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae not found: ID does not exist" containerID="f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.208007 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae"} err="failed to get container status \"f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae\": rpc error: code = NotFound desc = could not find container \"f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae\": container with ID starting with f480af923535dfb3d920f8e2a2f0bc0d8a6b7eb8d8acc3e83ce741fb0ad69fae not found: ID does not exist" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.210645 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb\") pod \"21448dfb-51d7-490e-bde9-c6195b66014f\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.210687 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config\") pod \"21448dfb-51d7-490e-bde9-c6195b66014f\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.210793 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc\") pod \"21448dfb-51d7-490e-bde9-c6195b66014f\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.210865 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb\") pod \"21448dfb-51d7-490e-bde9-c6195b66014f\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.210908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8r84\" (UniqueName: \"kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84\") pod \"21448dfb-51d7-490e-bde9-c6195b66014f\" (UID: \"21448dfb-51d7-490e-bde9-c6195b66014f\") " Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.216426 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84" (OuterVolumeSpecName: "kube-api-access-v8r84") pod "21448dfb-51d7-490e-bde9-c6195b66014f" (UID: "21448dfb-51d7-490e-bde9-c6195b66014f"). InnerVolumeSpecName "kube-api-access-v8r84". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.250120 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config" (OuterVolumeSpecName: "config") pod "21448dfb-51d7-490e-bde9-c6195b66014f" (UID: "21448dfb-51d7-490e-bde9-c6195b66014f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.252541 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21448dfb-51d7-490e-bde9-c6195b66014f" (UID: "21448dfb-51d7-490e-bde9-c6195b66014f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.254843 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21448dfb-51d7-490e-bde9-c6195b66014f" (UID: "21448dfb-51d7-490e-bde9-c6195b66014f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.264194 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21448dfb-51d7-490e-bde9-c6195b66014f" (UID: "21448dfb-51d7-490e-bde9-c6195b66014f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.312615 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.312665 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.312679 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8r84\" (UniqueName: \"kubernetes.io/projected/21448dfb-51d7-490e-bde9-c6195b66014f-kube-api-access-v8r84\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.312688 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.312698 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21448dfb-51d7-490e-bde9-c6195b66014f-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.511807 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.518349 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78bb75bfd7-t92lf"] Oct 06 10:13:41 crc kubenswrapper[4989]: I1006 10:13:41.948763 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" path="/var/lib/kubelet/pods/21448dfb-51d7-490e-bde9-c6195b66014f/volumes" Oct 06 10:14:06 crc kubenswrapper[4989]: I1006 10:14:06.473554 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:14:07 crc kubenswrapper[4989]: I1006 10:14:07.479541 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-849964bbfd-whdvn" Oct 06 10:14:20 crc kubenswrapper[4989]: I1006 10:14:20.907029 4989 scope.go:117] "RemoveContainer" containerID="067666b00022b9c722ada3f7ecf2a1b88212fd231758196c1d85dbce67768234" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.963022 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-68dl8"] Oct 06 10:14:28 crc kubenswrapper[4989]: E1006 10:14:28.966897 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="dnsmasq-dns" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.967014 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="dnsmasq-dns" Oct 06 10:14:28 crc kubenswrapper[4989]: E1006 10:14:28.967108 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="init" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.967174 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="init" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.967447 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="21448dfb-51d7-490e-bde9-c6195b66014f" containerName="dnsmasq-dns" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.968220 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:28 crc kubenswrapper[4989]: I1006 10:14:28.983502 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-68dl8"] Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.009398 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zp4m\" (UniqueName: \"kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m\") pod \"nova-api-db-create-68dl8\" (UID: \"4e2645a3-e847-4998-97b3-054835ab535a\") " pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.110998 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zp4m\" (UniqueName: \"kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m\") pod \"nova-api-db-create-68dl8\" (UID: \"4e2645a3-e847-4998-97b3-054835ab535a\") " pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.132536 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zp4m\" (UniqueName: \"kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m\") pod \"nova-api-db-create-68dl8\" (UID: \"4e2645a3-e847-4998-97b3-054835ab535a\") " pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.151359 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-ls8gg"] Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.152809 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.158977 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ls8gg"] Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.251604 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xxln2"] Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.252910 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.261666 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxln2"] Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.299094 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.314649 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5vtr\" (UniqueName: \"kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr\") pod \"nova-cell1-db-create-xxln2\" (UID: \"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5\") " pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.314917 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68gh\" (UniqueName: \"kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh\") pod \"nova-cell0-db-create-ls8gg\" (UID: \"300b916c-cba2-4226-b79c-d32d69454b49\") " pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.415912 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68gh\" (UniqueName: \"kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh\") pod \"nova-cell0-db-create-ls8gg\" (UID: \"300b916c-cba2-4226-b79c-d32d69454b49\") " pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.416170 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5vtr\" (UniqueName: \"kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr\") pod \"nova-cell1-db-create-xxln2\" (UID: \"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5\") " pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.445408 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68gh\" (UniqueName: \"kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh\") pod \"nova-cell0-db-create-ls8gg\" (UID: \"300b916c-cba2-4226-b79c-d32d69454b49\") " pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.466641 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5vtr\" (UniqueName: \"kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr\") pod \"nova-cell1-db-create-xxln2\" (UID: \"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5\") " pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.502819 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.568227 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:29 crc kubenswrapper[4989]: I1006 10:14:29.772024 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-68dl8"] Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.060484 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ls8gg"] Oct 06 10:14:30 crc kubenswrapper[4989]: W1006 10:14:30.065992 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod300b916c_cba2_4226_b79c_d32d69454b49.slice/crio-fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6 WatchSource:0}: Error finding container fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6: Status 404 returned error can't find the container with id fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6 Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.125422 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxln2"] Oct 06 10:14:30 crc kubenswrapper[4989]: W1006 10:14:30.130808 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode50c946b_ccfa_4b43_8e4f_8b47cdaaeed5.slice/crio-8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719 WatchSource:0}: Error finding container 8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719: Status 404 returned error can't find the container with id 8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719 Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.735276 4989 generic.go:334] "Generic (PLEG): container finished" podID="4e2645a3-e847-4998-97b3-054835ab535a" containerID="c425c248c742879d68ba3e729e6dfd0ce90e7dfe72f4b48b20ef03ac1daf3757" exitCode=0 Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.735332 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-68dl8" event={"ID":"4e2645a3-e847-4998-97b3-054835ab535a","Type":"ContainerDied","Data":"c425c248c742879d68ba3e729e6dfd0ce90e7dfe72f4b48b20ef03ac1daf3757"} Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.735396 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-68dl8" event={"ID":"4e2645a3-e847-4998-97b3-054835ab535a","Type":"ContainerStarted","Data":"92074a1e085f10ed27ccf8e37f7b8284e12a3c6c6bdc20adf8c22db1cea36ee5"} Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.737328 4989 generic.go:334] "Generic (PLEG): container finished" podID="e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" containerID="e4e9086cdf4328a23d6dbf6690f425192c84da003609ad354d32447d02120bf4" exitCode=0 Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.737358 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxln2" event={"ID":"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5","Type":"ContainerDied","Data":"e4e9086cdf4328a23d6dbf6690f425192c84da003609ad354d32447d02120bf4"} Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.737379 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxln2" event={"ID":"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5","Type":"ContainerStarted","Data":"8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719"} Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.739409 4989 generic.go:334] "Generic (PLEG): container finished" podID="300b916c-cba2-4226-b79c-d32d69454b49" containerID="bcaac616ade48e0614a4c1777146ba0c218b9b16a0465aa989b00c7369163796" exitCode=0 Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.739497 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ls8gg" event={"ID":"300b916c-cba2-4226-b79c-d32d69454b49","Type":"ContainerDied","Data":"bcaac616ade48e0614a4c1777146ba0c218b9b16a0465aa989b00c7369163796"} Oct 06 10:14:30 crc kubenswrapper[4989]: I1006 10:14:30.739568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ls8gg" event={"ID":"300b916c-cba2-4226-b79c-d32d69454b49","Type":"ContainerStarted","Data":"fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6"} Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.271794 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.276763 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.283866 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.371590 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zp4m\" (UniqueName: \"kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m\") pod \"4e2645a3-e847-4998-97b3-054835ab535a\" (UID: \"4e2645a3-e847-4998-97b3-054835ab535a\") " Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.371714 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w68gh\" (UniqueName: \"kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh\") pod \"300b916c-cba2-4226-b79c-d32d69454b49\" (UID: \"300b916c-cba2-4226-b79c-d32d69454b49\") " Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.371847 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5vtr\" (UniqueName: \"kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr\") pod \"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5\" (UID: \"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5\") " Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.377788 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m" (OuterVolumeSpecName: "kube-api-access-7zp4m") pod "4e2645a3-e847-4998-97b3-054835ab535a" (UID: "4e2645a3-e847-4998-97b3-054835ab535a"). InnerVolumeSpecName "kube-api-access-7zp4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.377963 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh" (OuterVolumeSpecName: "kube-api-access-w68gh") pod "300b916c-cba2-4226-b79c-d32d69454b49" (UID: "300b916c-cba2-4226-b79c-d32d69454b49"). InnerVolumeSpecName "kube-api-access-w68gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.379534 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr" (OuterVolumeSpecName: "kube-api-access-t5vtr") pod "e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" (UID: "e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5"). InnerVolumeSpecName "kube-api-access-t5vtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.473970 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zp4m\" (UniqueName: \"kubernetes.io/projected/4e2645a3-e847-4998-97b3-054835ab535a-kube-api-access-7zp4m\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.474023 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w68gh\" (UniqueName: \"kubernetes.io/projected/300b916c-cba2-4226-b79c-d32d69454b49-kube-api-access-w68gh\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.474043 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5vtr\" (UniqueName: \"kubernetes.io/projected/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5-kube-api-access-t5vtr\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.774371 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxln2" event={"ID":"e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5","Type":"ContainerDied","Data":"8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719"} Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.774413 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxln2" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.774437 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8521e411a5193e5f832ed18b0f5e582fd7a2e95fd6fab31d917530396812f719" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.777589 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ls8gg" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.777734 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ls8gg" event={"ID":"300b916c-cba2-4226-b79c-d32d69454b49","Type":"ContainerDied","Data":"fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6"} Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.777766 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fde3abc79a0fac5673bd4cd12143d965cb6d89e1fc7e7d9fc901d60343c981c6" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.780709 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-68dl8" event={"ID":"4e2645a3-e847-4998-97b3-054835ab535a","Type":"ContainerDied","Data":"92074a1e085f10ed27ccf8e37f7b8284e12a3c6c6bdc20adf8c22db1cea36ee5"} Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.780753 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92074a1e085f10ed27ccf8e37f7b8284e12a3c6c6bdc20adf8c22db1cea36ee5" Oct 06 10:14:32 crc kubenswrapper[4989]: I1006 10:14:32.780809 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-68dl8" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.210464 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f87a-account-create-vmjjt"] Oct 06 10:14:39 crc kubenswrapper[4989]: E1006 10:14:39.212050 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300b916c-cba2-4226-b79c-d32d69454b49" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212086 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="300b916c-cba2-4226-b79c-d32d69454b49" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: E1006 10:14:39.212115 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212132 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: E1006 10:14:39.212196 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2645a3-e847-4998-97b3-054835ab535a" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212213 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2645a3-e847-4998-97b3-054835ab535a" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212605 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="300b916c-cba2-4226-b79c-d32d69454b49" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212637 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.212688 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2645a3-e847-4998-97b3-054835ab535a" containerName="mariadb-database-create" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.213786 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.215733 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.226645 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f87a-account-create-vmjjt"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.307028 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frds\" (UniqueName: \"kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds\") pod \"nova-api-f87a-account-create-vmjjt\" (UID: \"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636\") " pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.404372 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a2dc-account-create-vjvqk"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.406400 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.408254 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frds\" (UniqueName: \"kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds\") pod \"nova-api-f87a-account-create-vmjjt\" (UID: \"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636\") " pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.408970 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.420046 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a2dc-account-create-vjvqk"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.435089 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frds\" (UniqueName: \"kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds\") pod \"nova-api-f87a-account-create-vmjjt\" (UID: \"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636\") " pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.509888 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnsc\" (UniqueName: \"kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc\") pod \"nova-cell0-a2dc-account-create-vjvqk\" (UID: \"0013c281-c9bb-4c43-b20c-4cd6cee09ebb\") " pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.552320 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.594037 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-40b9-account-create-7c2t2"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.595678 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.601256 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.605800 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-40b9-account-create-7c2t2"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.612381 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnsc\" (UniqueName: \"kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc\") pod \"nova-cell0-a2dc-account-create-vjvqk\" (UID: \"0013c281-c9bb-4c43-b20c-4cd6cee09ebb\") " pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.644203 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnsc\" (UniqueName: \"kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc\") pod \"nova-cell0-a2dc-account-create-vjvqk\" (UID: \"0013c281-c9bb-4c43-b20c-4cd6cee09ebb\") " pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.713856 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccfv8\" (UniqueName: \"kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8\") pod \"nova-cell1-40b9-account-create-7c2t2\" (UID: \"059c270b-9842-4228-a836-a256a05d570b\") " pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.817862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccfv8\" (UniqueName: \"kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8\") pod \"nova-cell1-40b9-account-create-7c2t2\" (UID: \"059c270b-9842-4228-a836-a256a05d570b\") " pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.821181 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.834336 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f87a-account-create-vmjjt"] Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.836061 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccfv8\" (UniqueName: \"kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8\") pod \"nova-cell1-40b9-account-create-7c2t2\" (UID: \"059c270b-9842-4228-a836-a256a05d570b\") " pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:39 crc kubenswrapper[4989]: I1006 10:14:39.992248 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.276506 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a2dc-account-create-vjvqk"] Oct 06 10:14:40 crc kubenswrapper[4989]: W1006 10:14:40.282798 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0013c281_c9bb_4c43_b20c_4cd6cee09ebb.slice/crio-572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067 WatchSource:0}: Error finding container 572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067: Status 404 returned error can't find the container with id 572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067 Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.459910 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-40b9-account-create-7c2t2"] Oct 06 10:14:40 crc kubenswrapper[4989]: W1006 10:14:40.469850 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod059c270b_9842_4228_a836_a256a05d570b.slice/crio-85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c WatchSource:0}: Error finding container 85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c: Status 404 returned error can't find the container with id 85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.879962 4989 generic.go:334] "Generic (PLEG): container finished" podID="059c270b-9842-4228-a836-a256a05d570b" containerID="0f9a5aa742eb800b5c99e6eabb0e3c6458f88647a51c7255f499e8e2477c95b0" exitCode=0 Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.880214 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-40b9-account-create-7c2t2" event={"ID":"059c270b-9842-4228-a836-a256a05d570b","Type":"ContainerDied","Data":"0f9a5aa742eb800b5c99e6eabb0e3c6458f88647a51c7255f499e8e2477c95b0"} Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.880266 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-40b9-account-create-7c2t2" event={"ID":"059c270b-9842-4228-a836-a256a05d570b","Type":"ContainerStarted","Data":"85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c"} Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.889586 4989 generic.go:334] "Generic (PLEG): container finished" podID="dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" containerID="01e69d9212f127c40568a4f89fe9b3961e9bbda94bde5151fe3355b118187f12" exitCode=0 Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.889719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f87a-account-create-vmjjt" event={"ID":"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636","Type":"ContainerDied","Data":"01e69d9212f127c40568a4f89fe9b3961e9bbda94bde5151fe3355b118187f12"} Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.889797 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f87a-account-create-vmjjt" event={"ID":"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636","Type":"ContainerStarted","Data":"d129e8c243cf0cf5f6ffe9b21867645b38e0a1b5f7df70c4497ef9884d0e6cca"} Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.892474 4989 generic.go:334] "Generic (PLEG): container finished" podID="0013c281-c9bb-4c43-b20c-4cd6cee09ebb" containerID="90932bd0f6d74676b81c1d46a91e4164bd30a0572ad080e0142a1d6f0ad26c23" exitCode=0 Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.892541 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" event={"ID":"0013c281-c9bb-4c43-b20c-4cd6cee09ebb","Type":"ContainerDied","Data":"90932bd0f6d74676b81c1d46a91e4164bd30a0572ad080e0142a1d6f0ad26c23"} Oct 06 10:14:40 crc kubenswrapper[4989]: I1006 10:14:40.892612 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" event={"ID":"0013c281-c9bb-4c43-b20c-4cd6cee09ebb","Type":"ContainerStarted","Data":"572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067"} Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.427217 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.435819 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.453566 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.568788 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccfv8\" (UniqueName: \"kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8\") pod \"059c270b-9842-4228-a836-a256a05d570b\" (UID: \"059c270b-9842-4228-a836-a256a05d570b\") " Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.569117 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4frds\" (UniqueName: \"kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds\") pod \"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636\" (UID: \"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636\") " Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.569212 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bnsc\" (UniqueName: \"kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc\") pod \"0013c281-c9bb-4c43-b20c-4cd6cee09ebb\" (UID: \"0013c281-c9bb-4c43-b20c-4cd6cee09ebb\") " Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.575030 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc" (OuterVolumeSpecName: "kube-api-access-7bnsc") pod "0013c281-c9bb-4c43-b20c-4cd6cee09ebb" (UID: "0013c281-c9bb-4c43-b20c-4cd6cee09ebb"). InnerVolumeSpecName "kube-api-access-7bnsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.575695 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds" (OuterVolumeSpecName: "kube-api-access-4frds") pod "dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" (UID: "dbaef9b1-96ff-4afc-a1f9-49fb7ae41636"). InnerVolumeSpecName "kube-api-access-4frds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.579831 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8" (OuterVolumeSpecName: "kube-api-access-ccfv8") pod "059c270b-9842-4228-a836-a256a05d570b" (UID: "059c270b-9842-4228-a836-a256a05d570b"). InnerVolumeSpecName "kube-api-access-ccfv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.671193 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4frds\" (UniqueName: \"kubernetes.io/projected/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636-kube-api-access-4frds\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.671227 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bnsc\" (UniqueName: \"kubernetes.io/projected/0013c281-c9bb-4c43-b20c-4cd6cee09ebb-kube-api-access-7bnsc\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.671243 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccfv8\" (UniqueName: \"kubernetes.io/projected/059c270b-9842-4228-a836-a256a05d570b-kube-api-access-ccfv8\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.910637 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-40b9-account-create-7c2t2" event={"ID":"059c270b-9842-4228-a836-a256a05d570b","Type":"ContainerDied","Data":"85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c"} Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.910751 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85acc6872d880da44d1be9b5e789c6a3fa8e441c820037b5307299ee9b9d7c0c" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.910698 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-40b9-account-create-7c2t2" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.913196 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f87a-account-create-vmjjt" event={"ID":"dbaef9b1-96ff-4afc-a1f9-49fb7ae41636","Type":"ContainerDied","Data":"d129e8c243cf0cf5f6ffe9b21867645b38e0a1b5f7df70c4497ef9884d0e6cca"} Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.913242 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d129e8c243cf0cf5f6ffe9b21867645b38e0a1b5f7df70c4497ef9884d0e6cca" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.913263 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f87a-account-create-vmjjt" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.915424 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" event={"ID":"0013c281-c9bb-4c43-b20c-4cd6cee09ebb","Type":"ContainerDied","Data":"572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067"} Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.915453 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="572760d1f13132024f72226b7408d688aaee5242de474b111f169913d904b067" Oct 06 10:14:42 crc kubenswrapper[4989]: I1006 10:14:42.915515 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a2dc-account-create-vjvqk" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.763012 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gc6mj"] Oct 06 10:14:44 crc kubenswrapper[4989]: E1006 10:14:44.764498 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0013c281-c9bb-4c43-b20c-4cd6cee09ebb" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.764520 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0013c281-c9bb-4c43-b20c-4cd6cee09ebb" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: E1006 10:14:44.764570 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.764582 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: E1006 10:14:44.764622 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059c270b-9842-4228-a836-a256a05d570b" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.764635 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="059c270b-9842-4228-a836-a256a05d570b" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.767365 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="059c270b-9842-4228-a836-a256a05d570b" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.767404 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0013c281-c9bb-4c43-b20c-4cd6cee09ebb" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.767440 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" containerName="mariadb-account-create" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.768870 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.786914 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2kp5r" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.787132 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.791249 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.816374 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gc6mj"] Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.918238 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.918303 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzdcg\" (UniqueName: \"kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.918543 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:44 crc kubenswrapper[4989]: I1006 10:14:44.918710 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.021155 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.021338 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzdcg\" (UniqueName: \"kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.021523 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.021702 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.028472 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.028595 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.029914 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.041510 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzdcg\" (UniqueName: \"kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg\") pod \"nova-cell0-conductor-db-sync-gc6mj\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.136324 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.459776 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gc6mj"] Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.952815 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" event={"ID":"c3958753-2e0a-47be-b2c1-5deaa2920c85","Type":"ContainerStarted","Data":"8c8739101f7578b22ef1571b78316d6e773b0f3122431cb0ca7d7ed25acdc41d"} Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.953168 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" event={"ID":"c3958753-2e0a-47be-b2c1-5deaa2920c85","Type":"ContainerStarted","Data":"963597e8b08523b83da5579d81d3263c9869577d63dcb2fef69f4fa705b4c5ac"} Oct 06 10:14:45 crc kubenswrapper[4989]: I1006 10:14:45.983507 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" podStartSLOduration=1.983484624 podStartE2EDuration="1.983484624s" podCreationTimestamp="2025-10-06 10:14:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:14:45.971974083 +0000 UTC m=+5736.761999673" watchObservedRunningTime="2025-10-06 10:14:45.983484624 +0000 UTC m=+5736.773510224" Oct 06 10:14:51 crc kubenswrapper[4989]: I1006 10:14:51.021159 4989 generic.go:334] "Generic (PLEG): container finished" podID="c3958753-2e0a-47be-b2c1-5deaa2920c85" containerID="8c8739101f7578b22ef1571b78316d6e773b0f3122431cb0ca7d7ed25acdc41d" exitCode=0 Oct 06 10:14:51 crc kubenswrapper[4989]: I1006 10:14:51.021248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" event={"ID":"c3958753-2e0a-47be-b2c1-5deaa2920c85","Type":"ContainerDied","Data":"8c8739101f7578b22ef1571b78316d6e773b0f3122431cb0ca7d7ed25acdc41d"} Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.478450 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.561861 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data\") pod \"c3958753-2e0a-47be-b2c1-5deaa2920c85\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.562111 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle\") pod \"c3958753-2e0a-47be-b2c1-5deaa2920c85\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.562169 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts\") pod \"c3958753-2e0a-47be-b2c1-5deaa2920c85\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.562347 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzdcg\" (UniqueName: \"kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg\") pod \"c3958753-2e0a-47be-b2c1-5deaa2920c85\" (UID: \"c3958753-2e0a-47be-b2c1-5deaa2920c85\") " Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.572438 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts" (OuterVolumeSpecName: "scripts") pod "c3958753-2e0a-47be-b2c1-5deaa2920c85" (UID: "c3958753-2e0a-47be-b2c1-5deaa2920c85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.579224 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg" (OuterVolumeSpecName: "kube-api-access-hzdcg") pod "c3958753-2e0a-47be-b2c1-5deaa2920c85" (UID: "c3958753-2e0a-47be-b2c1-5deaa2920c85"). InnerVolumeSpecName "kube-api-access-hzdcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.599856 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3958753-2e0a-47be-b2c1-5deaa2920c85" (UID: "c3958753-2e0a-47be-b2c1-5deaa2920c85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.617820 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data" (OuterVolumeSpecName: "config-data") pod "c3958753-2e0a-47be-b2c1-5deaa2920c85" (UID: "c3958753-2e0a-47be-b2c1-5deaa2920c85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.666153 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.666239 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.666260 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzdcg\" (UniqueName: \"kubernetes.io/projected/c3958753-2e0a-47be-b2c1-5deaa2920c85-kube-api-access-hzdcg\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:52 crc kubenswrapper[4989]: I1006 10:14:52.666278 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3958753-2e0a-47be-b2c1-5deaa2920c85-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.048169 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" event={"ID":"c3958753-2e0a-47be-b2c1-5deaa2920c85","Type":"ContainerDied","Data":"963597e8b08523b83da5579d81d3263c9869577d63dcb2fef69f4fa705b4c5ac"} Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.048215 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="963597e8b08523b83da5579d81d3263c9869577d63dcb2fef69f4fa705b4c5ac" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.048285 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gc6mj" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.152634 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:14:53 crc kubenswrapper[4989]: E1006 10:14:53.153096 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3958753-2e0a-47be-b2c1-5deaa2920c85" containerName="nova-cell0-conductor-db-sync" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.153119 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3958753-2e0a-47be-b2c1-5deaa2920c85" containerName="nova-cell0-conductor-db-sync" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.153347 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3958753-2e0a-47be-b2c1-5deaa2920c85" containerName="nova-cell0-conductor-db-sync" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.157169 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.159560 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2kp5r" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.165507 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.166018 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.175491 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.175597 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96qpc\" (UniqueName: \"kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.175690 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.279308 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96qpc\" (UniqueName: \"kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.279379 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.279695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.286419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.293524 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.314409 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96qpc\" (UniqueName: \"kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc\") pod \"nova-cell0-conductor-0\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:53 crc kubenswrapper[4989]: I1006 10:14:53.481867 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:54 crc kubenswrapper[4989]: I1006 10:14:54.020922 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:14:54 crc kubenswrapper[4989]: I1006 10:14:54.059602 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"526b003d-1c65-453c-9c1f-31003fa6d0a3","Type":"ContainerStarted","Data":"cf8c0d7e2c619b04960c55b0b459ca223b194c83fb1033267f3b2ce29145a771"} Oct 06 10:14:55 crc kubenswrapper[4989]: I1006 10:14:55.072323 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"526b003d-1c65-453c-9c1f-31003fa6d0a3","Type":"ContainerStarted","Data":"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4"} Oct 06 10:14:55 crc kubenswrapper[4989]: I1006 10:14:55.072958 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 06 10:14:55 crc kubenswrapper[4989]: I1006 10:14:55.092175 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.092075337 podStartE2EDuration="2.092075337s" podCreationTimestamp="2025-10-06 10:14:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:14:55.084268382 +0000 UTC m=+5745.874293962" watchObservedRunningTime="2025-10-06 10:14:55.092075337 +0000 UTC m=+5745.882100927" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.166840 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx"] Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.174147 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.177308 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.178842 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.179629 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx"] Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.222702 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.222797 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttdv8\" (UniqueName: \"kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.222838 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.323694 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.323778 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttdv8\" (UniqueName: \"kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.323821 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.325141 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.336042 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.354241 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttdv8\" (UniqueName: \"kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8\") pod \"collect-profiles-29329095-xf7gx\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.508357 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:00 crc kubenswrapper[4989]: W1006 10:15:00.967837 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8c6e7a4_1259_43c2_bd73_e7daa82f34b5.slice/crio-daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d WatchSource:0}: Error finding container daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d: Status 404 returned error can't find the container with id daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d Oct 06 10:15:00 crc kubenswrapper[4989]: I1006 10:15:00.971883 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx"] Oct 06 10:15:01 crc kubenswrapper[4989]: I1006 10:15:01.136005 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" event={"ID":"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5","Type":"ContainerStarted","Data":"daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d"} Oct 06 10:15:02 crc kubenswrapper[4989]: I1006 10:15:02.149310 4989 generic.go:334] "Generic (PLEG): container finished" podID="b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" containerID="6b2b51b00ab758b5d651e1645075b3ed8c0bf450616374096fa7635dc4f4ca57" exitCode=0 Oct 06 10:15:02 crc kubenswrapper[4989]: I1006 10:15:02.149442 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" event={"ID":"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5","Type":"ContainerDied","Data":"6b2b51b00ab758b5d651e1645075b3ed8c0bf450616374096fa7635dc4f4ca57"} Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.518494 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.607091 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.797136 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttdv8\" (UniqueName: \"kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8\") pod \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.797350 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume\") pod \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.797478 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume\") pod \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\" (UID: \"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5\") " Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.798519 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume" (OuterVolumeSpecName: "config-volume") pod "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" (UID: "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.805018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8" (OuterVolumeSpecName: "kube-api-access-ttdv8") pod "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" (UID: "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5"). InnerVolumeSpecName "kube-api-access-ttdv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.805177 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" (UID: "b8c6e7a4-1259-43c2-bd73-e7daa82f34b5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.900822 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.900889 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttdv8\" (UniqueName: \"kubernetes.io/projected/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-kube-api-access-ttdv8\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.900909 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.937190 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:15:03 crc kubenswrapper[4989]: I1006 10:15:03.937266 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.065688 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-66442"] Oct 06 10:15:04 crc kubenswrapper[4989]: E1006 10:15:04.066100 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.066119 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.066336 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" containerName="collect-profiles" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.067066 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.069705 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.069933 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.086179 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-66442"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.165514 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" event={"ID":"b8c6e7a4-1259-43c2-bd73-e7daa82f34b5","Type":"ContainerDied","Data":"daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d"} Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.165552 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daed8023dbafeeb6a37d472184c8e399e673a8940d06d3bea42b385c471b2a9d" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.165581 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.174257 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.175870 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.180296 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.194723 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.206183 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.206549 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.206705 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.206832 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sm6g\" (UniqueName: \"kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.258086 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.259124 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.261530 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.275897 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.308727 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.308963 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309052 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309138 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sm6g\" (UniqueName: \"kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309221 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kngg7\" (UniqueName: \"kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309300 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309390 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.309552 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.313162 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.313381 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.314381 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.342439 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sm6g\" (UniqueName: \"kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g\") pod \"nova-cell0-cell-mapping-66442\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.385135 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.389785 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.391329 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.395872 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.402458 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.403908 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.411710 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.411853 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.411956 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kngg7\" (UniqueName: \"kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.411975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.412001 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbqnf\" (UniqueName: \"kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.412073 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.412107 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.412156 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.417334 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.421434 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.422882 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.440305 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kngg7\" (UniqueName: \"kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.449808 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.452519 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.495094 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.496464 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.508213 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516487 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516540 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516613 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516641 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516689 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516729 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n48zr\" (UniqueName: \"kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516782 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbqnf\" (UniqueName: \"kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516842 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-247ml\" (UniqueName: \"kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.516867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.537339 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.544409 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.546121 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.554512 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbqnf\" (UniqueName: \"kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf\") pod \"nova-scheduler-0\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618039 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfgh7\" (UniqueName: \"kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618081 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618105 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-247ml\" (UniqueName: \"kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618200 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618221 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618254 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618293 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618320 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.618351 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n48zr\" (UniqueName: \"kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.621030 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.623196 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.623605 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.631748 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.647148 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.650158 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n48zr\" (UniqueName: \"kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr\") pod \"nova-cell1-novncproxy-0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.663125 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-247ml\" (UniqueName: \"kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml\") pod \"nova-metadata-0\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.671880 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.696672 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.703637 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329050-prdpd"] Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.708981 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.720228 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.720270 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.720428 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfgh7\" (UniqueName: \"kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.720454 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.720484 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.721213 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.721443 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.721967 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.724357 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.739971 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfgh7\" (UniqueName: \"kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7\") pod \"dnsmasq-dns-dcf6b7bb9-h75b4\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.750580 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:04 crc kubenswrapper[4989]: I1006 10:15:04.936275 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.053107 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-66442"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.066241 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.121037 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.176836 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3adf6a92-096f-4284-af10-9f703ff47b95","Type":"ContainerStarted","Data":"fd336bafde9e0624ac0975079a501058beb3994c2e79c8e20c0ae40ebf503772"} Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.199984 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerStarted","Data":"25c858b8936ce81744beacd23df86c6ed0145eda9a3ed63e64c6c91b27b6b32a"} Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.202727 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-66442" event={"ID":"f7e54d59-17a2-46bc-8188-51c1d67b82d8","Type":"ContainerStarted","Data":"2de3d5c72eb67fd7e4434d0ae05e600b792259fd9a06acdb16bcac29c5e294bc"} Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.476256 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:05 crc kubenswrapper[4989]: W1006 10:15:05.476939 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c0e7098_5c8c_4363_b1b7_01c10b695fc1.slice/crio-6ed28c8220fe523dcea1aa443b453f8e6b1376eb66e8caf0693aaeac58f03ef0 WatchSource:0}: Error finding container 6ed28c8220fe523dcea1aa443b453f8e6b1376eb66e8caf0693aaeac58f03ef0: Status 404 returned error can't find the container with id 6ed28c8220fe523dcea1aa443b453f8e6b1376eb66e8caf0693aaeac58f03ef0 Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.519710 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.528004 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.644689 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nwvvc"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.645992 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.648471 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.662215 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.686878 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nwvvc"] Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.748335 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.748789 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.748895 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.748915 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n7n4\" (UniqueName: \"kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.850808 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.850902 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.850959 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n7n4\" (UniqueName: \"kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.850978 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.857398 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.857853 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.857999 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.881612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n7n4\" (UniqueName: \"kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4\") pod \"nova-cell1-conductor-db-sync-nwvvc\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.902946 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:05 crc kubenswrapper[4989]: I1006 10:15:05.957552 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e3ed952-b113-406d-a267-22445a1e4ead" path="/var/lib/kubelet/pods/6e3ed952-b113-406d-a267-22445a1e4ead/volumes" Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.216687 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-66442" event={"ID":"f7e54d59-17a2-46bc-8188-51c1d67b82d8","Type":"ContainerStarted","Data":"26c8ef0b6ca527dc9bafc25e08579ed1053186c26a778001e06bbd575764f951"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.225068 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3adf6a92-096f-4284-af10-9f703ff47b95","Type":"ContainerStarted","Data":"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.227038 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerStarted","Data":"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.227081 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerStarted","Data":"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.232084 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerStarted","Data":"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.232126 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerStarted","Data":"6ed28c8220fe523dcea1aa443b453f8e6b1376eb66e8caf0693aaeac58f03ef0"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.235699 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1b4c62a-3cae-4649-b041-91fc4cd606f0","Type":"ContainerStarted","Data":"66dbc734a43dd45eccc68b9de1126961a0923b2713543855cbd7bc540572f70e"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.235736 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1b4c62a-3cae-4649-b041-91fc4cd606f0","Type":"ContainerStarted","Data":"af4936eaeca81ee6358c02896e480191c567bb48f8b9598074b4f80924a58966"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.238153 4989 generic.go:334] "Generic (PLEG): container finished" podID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerID="8afb79fc1865988b10e85b358785958e108891022b2b1c9dfab3860ef75bbf8e" exitCode=0 Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.238198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" event={"ID":"9cce2f3f-e874-4c9d-805f-784219fd05d5","Type":"ContainerDied","Data":"8afb79fc1865988b10e85b358785958e108891022b2b1c9dfab3860ef75bbf8e"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.238223 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" event={"ID":"9cce2f3f-e874-4c9d-805f-784219fd05d5","Type":"ContainerStarted","Data":"b915c0cf5a3d1a7e419845f23011a9d60119b7eb81d8b704f9f2cca5be569bd7"} Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.245561 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-66442" podStartSLOduration=2.245540778 podStartE2EDuration="2.245540778s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:06.237668862 +0000 UTC m=+5757.027694452" watchObservedRunningTime="2025-10-06 10:15:06.245540778 +0000 UTC m=+5757.035566358" Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.257092 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.25705147 podStartE2EDuration="2.25705147s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:06.252141748 +0000 UTC m=+5757.042167328" watchObservedRunningTime="2025-10-06 10:15:06.25705147 +0000 UTC m=+5757.047077050" Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.298356 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.298328578 podStartE2EDuration="2.298328578s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:06.271702451 +0000 UTC m=+5757.061728031" watchObservedRunningTime="2025-10-06 10:15:06.298328578 +0000 UTC m=+5757.088354158" Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.318827 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.318805727 podStartE2EDuration="2.318805727s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:06.288862985 +0000 UTC m=+5757.078888585" watchObservedRunningTime="2025-10-06 10:15:06.318805727 +0000 UTC m=+5757.108831307" Oct 06 10:15:06 crc kubenswrapper[4989]: I1006 10:15:06.417492 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nwvvc"] Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.251027 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" event={"ID":"9cce2f3f-e874-4c9d-805f-784219fd05d5","Type":"ContainerStarted","Data":"4b7850684d23fdd264dd1b306607c85ba0abcb31b67d45fec01245852f5b575b"} Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.251352 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.254949 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" event={"ID":"349115de-5839-4178-9af2-db9765b42d09","Type":"ContainerStarted","Data":"5ec1e4ec40d647079cc1dcb153c1fdee44b84ecabf63487745d1ae9c2c207227"} Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.255017 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" event={"ID":"349115de-5839-4178-9af2-db9765b42d09","Type":"ContainerStarted","Data":"7b40bcfaf4ba3fee7690492c14389ca6ef7eab99cbbc34e04d0fe7f1c56491fc"} Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.257650 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerStarted","Data":"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3"} Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.302552 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" podStartSLOduration=3.302531922 podStartE2EDuration="3.302531922s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:07.278232033 +0000 UTC m=+5758.068257653" watchObservedRunningTime="2025-10-06 10:15:07.302531922 +0000 UTC m=+5758.092557522" Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.305158 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" podStartSLOduration=2.3051489370000002 podStartE2EDuration="2.305148937s" podCreationTimestamp="2025-10-06 10:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:07.295100148 +0000 UTC m=+5758.085125758" watchObservedRunningTime="2025-10-06 10:15:07.305148937 +0000 UTC m=+5758.095174527" Oct 06 10:15:07 crc kubenswrapper[4989]: I1006 10:15:07.320383 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.320358655 podStartE2EDuration="3.320358655s" podCreationTimestamp="2025-10-06 10:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:07.312212371 +0000 UTC m=+5758.102237971" watchObservedRunningTime="2025-10-06 10:15:07.320358655 +0000 UTC m=+5758.110384245" Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.277139 4989 generic.go:334] "Generic (PLEG): container finished" podID="349115de-5839-4178-9af2-db9765b42d09" containerID="5ec1e4ec40d647079cc1dcb153c1fdee44b84ecabf63487745d1ae9c2c207227" exitCode=0 Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.277234 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" event={"ID":"349115de-5839-4178-9af2-db9765b42d09","Type":"ContainerDied","Data":"5ec1e4ec40d647079cc1dcb153c1fdee44b84ecabf63487745d1ae9c2c207227"} Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.672537 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.709531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.750759 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:09 crc kubenswrapper[4989]: I1006 10:15:09.750841 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.306829 4989 generic.go:334] "Generic (PLEG): container finished" podID="f7e54d59-17a2-46bc-8188-51c1d67b82d8" containerID="26c8ef0b6ca527dc9bafc25e08579ed1053186c26a778001e06bbd575764f951" exitCode=0 Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.307384 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-66442" event={"ID":"f7e54d59-17a2-46bc-8188-51c1d67b82d8","Type":"ContainerDied","Data":"26c8ef0b6ca527dc9bafc25e08579ed1053186c26a778001e06bbd575764f951"} Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.663921 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.762315 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n7n4\" (UniqueName: \"kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4\") pod \"349115de-5839-4178-9af2-db9765b42d09\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.762395 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle\") pod \"349115de-5839-4178-9af2-db9765b42d09\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.762481 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data\") pod \"349115de-5839-4178-9af2-db9765b42d09\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.762522 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts\") pod \"349115de-5839-4178-9af2-db9765b42d09\" (UID: \"349115de-5839-4178-9af2-db9765b42d09\") " Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.768453 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts" (OuterVolumeSpecName: "scripts") pod "349115de-5839-4178-9af2-db9765b42d09" (UID: "349115de-5839-4178-9af2-db9765b42d09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.769033 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4" (OuterVolumeSpecName: "kube-api-access-8n7n4") pod "349115de-5839-4178-9af2-db9765b42d09" (UID: "349115de-5839-4178-9af2-db9765b42d09"). InnerVolumeSpecName "kube-api-access-8n7n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.791305 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "349115de-5839-4178-9af2-db9765b42d09" (UID: "349115de-5839-4178-9af2-db9765b42d09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.794055 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data" (OuterVolumeSpecName: "config-data") pod "349115de-5839-4178-9af2-db9765b42d09" (UID: "349115de-5839-4178-9af2-db9765b42d09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.863982 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.864014 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.864023 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n7n4\" (UniqueName: \"kubernetes.io/projected/349115de-5839-4178-9af2-db9765b42d09-kube-api-access-8n7n4\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:10 crc kubenswrapper[4989]: I1006 10:15:10.864033 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/349115de-5839-4178-9af2-db9765b42d09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.323260 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" event={"ID":"349115de-5839-4178-9af2-db9765b42d09","Type":"ContainerDied","Data":"7b40bcfaf4ba3fee7690492c14389ca6ef7eab99cbbc34e04d0fe7f1c56491fc"} Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.323342 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-nwvvc" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.323356 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b40bcfaf4ba3fee7690492c14389ca6ef7eab99cbbc34e04d0fe7f1c56491fc" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.766672 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:15:11 crc kubenswrapper[4989]: E1006 10:15:11.767389 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349115de-5839-4178-9af2-db9765b42d09" containerName="nova-cell1-conductor-db-sync" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.767410 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="349115de-5839-4178-9af2-db9765b42d09" containerName="nova-cell1-conductor-db-sync" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.767641 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="349115de-5839-4178-9af2-db9765b42d09" containerName="nova-cell1-conductor-db-sync" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.768410 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.772337 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.775158 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.784776 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.883755 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sm6g\" (UniqueName: \"kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g\") pod \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.883861 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data\") pod \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.883975 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle\") pod \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.884044 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts\") pod \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\" (UID: \"f7e54d59-17a2-46bc-8188-51c1d67b82d8\") " Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.884250 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtsng\" (UniqueName: \"kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.884308 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.884370 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.889086 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g" (OuterVolumeSpecName: "kube-api-access-5sm6g") pod "f7e54d59-17a2-46bc-8188-51c1d67b82d8" (UID: "f7e54d59-17a2-46bc-8188-51c1d67b82d8"). InnerVolumeSpecName "kube-api-access-5sm6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.890580 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts" (OuterVolumeSpecName: "scripts") pod "f7e54d59-17a2-46bc-8188-51c1d67b82d8" (UID: "f7e54d59-17a2-46bc-8188-51c1d67b82d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.908573 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data" (OuterVolumeSpecName: "config-data") pod "f7e54d59-17a2-46bc-8188-51c1d67b82d8" (UID: "f7e54d59-17a2-46bc-8188-51c1d67b82d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.909090 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7e54d59-17a2-46bc-8188-51c1d67b82d8" (UID: "f7e54d59-17a2-46bc-8188-51c1d67b82d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.985676 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.985824 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.985964 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtsng\" (UniqueName: \"kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.986078 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.986121 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.986134 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e54d59-17a2-46bc-8188-51c1d67b82d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.986145 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sm6g\" (UniqueName: \"kubernetes.io/projected/f7e54d59-17a2-46bc-8188-51c1d67b82d8-kube-api-access-5sm6g\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.990246 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:11 crc kubenswrapper[4989]: I1006 10:15:11.990303 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.007196 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtsng\" (UniqueName: \"kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng\") pod \"nova-cell1-conductor-0\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.088033 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:12 crc kubenswrapper[4989]: E1006 10:15:12.105998 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7e54d59_17a2_46bc_8188_51c1d67b82d8.slice/crio-2de3d5c72eb67fd7e4434d0ae05e600b792259fd9a06acdb16bcac29c5e294bc\": RecentStats: unable to find data in memory cache]" Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.334487 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-66442" event={"ID":"f7e54d59-17a2-46bc-8188-51c1d67b82d8","Type":"ContainerDied","Data":"2de3d5c72eb67fd7e4434d0ae05e600b792259fd9a06acdb16bcac29c5e294bc"} Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.334837 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2de3d5c72eb67fd7e4434d0ae05e600b792259fd9a06acdb16bcac29c5e294bc" Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.334575 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-66442" Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.537880 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.538111 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-log" containerID="cri-o://d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" gracePeriod=30 Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.538583 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-api" containerID="cri-o://33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" gracePeriod=30 Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.558104 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.558322 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3adf6a92-096f-4284-af10-9f703ff47b95" containerName="nova-scheduler-scheduler" containerID="cri-o://30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f" gracePeriod=30 Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.601498 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:15:12 crc kubenswrapper[4989]: W1006 10:15:12.602033 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod557be752_c1ce_49a2_8cc0_7f73305c66b5.slice/crio-f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784 WatchSource:0}: Error finding container f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784: Status 404 returned error can't find the container with id f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784 Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.646426 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.647032 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-log" containerID="cri-o://b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" gracePeriod=30 Oct 06 10:15:12 crc kubenswrapper[4989]: I1006 10:15:12.647148 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-metadata" containerID="cri-o://ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" gracePeriod=30 Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.062796 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.144331 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.217124 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle\") pod \"19a56b63-c809-433c-93e8-b44187aa280b\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.217224 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs\") pod \"19a56b63-c809-433c-93e8-b44187aa280b\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.217282 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data\") pod \"19a56b63-c809-433c-93e8-b44187aa280b\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.217434 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kngg7\" (UniqueName: \"kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7\") pod \"19a56b63-c809-433c-93e8-b44187aa280b\" (UID: \"19a56b63-c809-433c-93e8-b44187aa280b\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.217611 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs" (OuterVolumeSpecName: "logs") pod "19a56b63-c809-433c-93e8-b44187aa280b" (UID: "19a56b63-c809-433c-93e8-b44187aa280b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.218169 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19a56b63-c809-433c-93e8-b44187aa280b-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.221854 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7" (OuterVolumeSpecName: "kube-api-access-kngg7") pod "19a56b63-c809-433c-93e8-b44187aa280b" (UID: "19a56b63-c809-433c-93e8-b44187aa280b"). InnerVolumeSpecName "kube-api-access-kngg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.240814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19a56b63-c809-433c-93e8-b44187aa280b" (UID: "19a56b63-c809-433c-93e8-b44187aa280b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.242629 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data" (OuterVolumeSpecName: "config-data") pod "19a56b63-c809-433c-93e8-b44187aa280b" (UID: "19a56b63-c809-433c-93e8-b44187aa280b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.319050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-247ml\" (UniqueName: \"kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml\") pod \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.319452 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle\") pod \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.319639 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data\") pod \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.319696 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs\") pod \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\" (UID: \"8c0e7098-5c8c-4363-b1b7-01c10b695fc1\") " Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.320098 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs" (OuterVolumeSpecName: "logs") pod "8c0e7098-5c8c-4363-b1b7-01c10b695fc1" (UID: "8c0e7098-5c8c-4363-b1b7-01c10b695fc1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.320203 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kngg7\" (UniqueName: \"kubernetes.io/projected/19a56b63-c809-433c-93e8-b44187aa280b-kube-api-access-kngg7\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.320263 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.320289 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19a56b63-c809-433c-93e8-b44187aa280b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.323011 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml" (OuterVolumeSpecName: "kube-api-access-247ml") pod "8c0e7098-5c8c-4363-b1b7-01c10b695fc1" (UID: "8c0e7098-5c8c-4363-b1b7-01c10b695fc1"). InnerVolumeSpecName "kube-api-access-247ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.350481 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c0e7098-5c8c-4363-b1b7-01c10b695fc1" (UID: "8c0e7098-5c8c-4363-b1b7-01c10b695fc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355424 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data" (OuterVolumeSpecName: "config-data") pod "8c0e7098-5c8c-4363-b1b7-01c10b695fc1" (UID: "8c0e7098-5c8c-4363-b1b7-01c10b695fc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355521 4989 generic.go:334] "Generic (PLEG): container finished" podID="19a56b63-c809-433c-93e8-b44187aa280b" containerID="33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" exitCode=0 Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355558 4989 generic.go:334] "Generic (PLEG): container finished" podID="19a56b63-c809-433c-93e8-b44187aa280b" containerID="d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" exitCode=143 Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355563 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerDied","Data":"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355624 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerDied","Data":"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355639 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"19a56b63-c809-433c-93e8-b44187aa280b","Type":"ContainerDied","Data":"25c858b8936ce81744beacd23df86c6ed0145eda9a3ed63e64c6c91b27b6b32a"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355670 4989 scope.go:117] "RemoveContainer" containerID="33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.355634 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359859 4989 generic.go:334] "Generic (PLEG): container finished" podID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerID="ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" exitCode=0 Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359893 4989 generic.go:334] "Generic (PLEG): container finished" podID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerID="b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" exitCode=143 Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359909 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359953 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerDied","Data":"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359983 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerDied","Data":"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.359996 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c0e7098-5c8c-4363-b1b7-01c10b695fc1","Type":"ContainerDied","Data":"6ed28c8220fe523dcea1aa443b453f8e6b1376eb66e8caf0693aaeac58f03ef0"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.368581 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"557be752-c1ce-49a2-8cc0-7f73305c66b5","Type":"ContainerStarted","Data":"9d75d55d504f174a69d3c97fb792d6cad40c23a296dc43d69448a5835f8d3c23"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.368639 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"557be752-c1ce-49a2-8cc0-7f73305c66b5","Type":"ContainerStarted","Data":"f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784"} Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.368955 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.396965 4989 scope.go:117] "RemoveContainer" containerID="d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.399673 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.3996393830000002 podStartE2EDuration="2.399639383s" podCreationTimestamp="2025-10-06 10:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:13.396467212 +0000 UTC m=+5764.186492792" watchObservedRunningTime="2025-10-06 10:15:13.399639383 +0000 UTC m=+5764.189664963" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.422419 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.422454 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.422466 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-247ml\" (UniqueName: \"kubernetes.io/projected/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-kube-api-access-247ml\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.422479 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c0e7098-5c8c-4363-b1b7-01c10b695fc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.423501 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.446719 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.446984 4989 scope.go:117] "RemoveContainer" containerID="33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.447449 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087\": container with ID starting with 33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087 not found: ID does not exist" containerID="33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.447486 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087"} err="failed to get container status \"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087\": rpc error: code = NotFound desc = could not find container \"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087\": container with ID starting with 33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.447513 4989 scope.go:117] "RemoveContainer" containerID="d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.449033 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032\": container with ID starting with d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032 not found: ID does not exist" containerID="d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.449062 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032"} err="failed to get container status \"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032\": rpc error: code = NotFound desc = could not find container \"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032\": container with ID starting with d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.449079 4989 scope.go:117] "RemoveContainer" containerID="33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.449325 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087"} err="failed to get container status \"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087\": rpc error: code = NotFound desc = could not find container \"33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087\": container with ID starting with 33026568997bd7f944383546d12d0182525b39aa3fbe3e7605b78a72e5349087 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.449349 4989 scope.go:117] "RemoveContainer" containerID="d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.450720 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032"} err="failed to get container status \"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032\": rpc error: code = NotFound desc = could not find container \"d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032\": container with ID starting with d6f137e9db70bc592ab7db79b6d22130e5295984ff4df95a7ec45e4cfbcbd032 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.450749 4989 scope.go:117] "RemoveContainer" containerID="ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.477981 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.484409 4989 scope.go:117] "RemoveContainer" containerID="b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.497141 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516032 4989 scope.go:117] "RemoveContainer" containerID="ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516261 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516604 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-metadata" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516620 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-metadata" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516630 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e54d59-17a2-46bc-8188-51c1d67b82d8" containerName="nova-manage" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516636 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e54d59-17a2-46bc-8188-51c1d67b82d8" containerName="nova-manage" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516670 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-log" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516679 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-log" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516689 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-log" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516696 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-log" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516716 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-api" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516722 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-api" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.516858 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3\": container with ID starting with ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3 not found: ID does not exist" containerID="ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516883 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-log" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516899 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a56b63-c809-433c-93e8-b44187aa280b" containerName="nova-api-api" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516911 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-log" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516922 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e54d59-17a2-46bc-8188-51c1d67b82d8" containerName="nova-manage" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516932 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" containerName="nova-metadata-metadata" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.516896 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3"} err="failed to get container status \"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3\": rpc error: code = NotFound desc = could not find container \"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3\": container with ID starting with ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.517011 4989 scope.go:117] "RemoveContainer" containerID="b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" Oct 06 10:15:13 crc kubenswrapper[4989]: E1006 10:15:13.517432 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770\": container with ID starting with b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770 not found: ID does not exist" containerID="b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.517464 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770"} err="failed to get container status \"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770\": rpc error: code = NotFound desc = could not find container \"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770\": container with ID starting with b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.517483 4989 scope.go:117] "RemoveContainer" containerID="ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.517738 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3"} err="failed to get container status \"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3\": rpc error: code = NotFound desc = could not find container \"ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3\": container with ID starting with ffda2d1b1229040349539504b48e10edcdf85ee9d50bc8300796e1b4ae1684e3 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.517762 4989 scope.go:117] "RemoveContainer" containerID="b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.518045 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.518044 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770"} err="failed to get container status \"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770\": rpc error: code = NotFound desc = could not find container \"b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770\": container with ID starting with b5ea9566c4b32c3b0e7bfc331dafaf57a78a34d82ffe32eceae79bcb0e03b770 not found: ID does not exist" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.520427 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.526527 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.541699 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.543562 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.549909 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.551837 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625276 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdcr2\" (UniqueName: \"kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625305 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hcg2\" (UniqueName: \"kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625408 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625537 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625569 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.625852 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.727426 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728523 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728606 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdcr2\" (UniqueName: \"kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728650 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hcg2\" (UniqueName: \"kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728783 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728819 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.728887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.729488 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.729755 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.735710 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.736121 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.736643 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.744765 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.748636 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdcr2\" (UniqueName: \"kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2\") pod \"nova-metadata-0\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.752296 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hcg2\" (UniqueName: \"kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2\") pod \"nova-api-0\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.837268 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.862537 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.961852 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a56b63-c809-433c-93e8-b44187aa280b" path="/var/lib/kubelet/pods/19a56b63-c809-433c-93e8-b44187aa280b/volumes" Oct 06 10:15:13 crc kubenswrapper[4989]: I1006 10:15:13.962519 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0e7098-5c8c-4363-b1b7-01c10b695fc1" path="/var/lib/kubelet/pods/8c0e7098-5c8c-4363-b1b7-01c10b695fc1/volumes" Oct 06 10:15:14 crc kubenswrapper[4989]: W1006 10:15:14.360054 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb1065a0_9d18_4c3f_af58_8a1a99948144.slice/crio-c5fab9b63d52a8a9fb29d5975a5dbb1b14d8d1ab78ca96fb9ed74703652fdb0d WatchSource:0}: Error finding container c5fab9b63d52a8a9fb29d5975a5dbb1b14d8d1ab78ca96fb9ed74703652fdb0d: Status 404 returned error can't find the container with id c5fab9b63d52a8a9fb29d5975a5dbb1b14d8d1ab78ca96fb9ed74703652fdb0d Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.360584 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.400630 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerStarted","Data":"c5fab9b63d52a8a9fb29d5975a5dbb1b14d8d1ab78ca96fb9ed74703652fdb0d"} Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.460508 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.710196 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.723417 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:14 crc kubenswrapper[4989]: I1006 10:15:14.938836 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.009894 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.010130 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="dnsmasq-dns" containerID="cri-o://0a4bb2572669056cfaa0c535b365a11c343e72325a6e6741efd32ebcdaa35195" gracePeriod=10 Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.411344 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerStarted","Data":"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.411705 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerStarted","Data":"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.414723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerStarted","Data":"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.414777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerStarted","Data":"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.414789 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerStarted","Data":"d8bd3410e22a5fbd594a3142a0fa3deaff61e32f70a4dd0a776f1300a15e85c2"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.416731 4989 generic.go:334] "Generic (PLEG): container finished" podID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerID="0a4bb2572669056cfaa0c535b365a11c343e72325a6e6741efd32ebcdaa35195" exitCode=0 Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.416766 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerDied","Data":"0a4bb2572669056cfaa0c535b365a11c343e72325a6e6741efd32ebcdaa35195"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.416791 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" event={"ID":"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5","Type":"ContainerDied","Data":"3dd0d9294036145712497635b9af000ac81ae8a1d79021e46f984db5fd38f192"} Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.416803 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dd0d9294036145712497635b9af000ac81ae8a1d79021e46f984db5fd38f192" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.431808 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.432092 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.432079364 podStartE2EDuration="2.432079364s" podCreationTimestamp="2025-10-06 10:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:15.428433389 +0000 UTC m=+5766.218458979" watchObservedRunningTime="2025-10-06 10:15:15.432079364 +0000 UTC m=+5766.222104944" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.483979 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.483958207 podStartE2EDuration="2.483958207s" podCreationTimestamp="2025-10-06 10:15:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:15.478249213 +0000 UTC m=+5766.268274803" watchObservedRunningTime="2025-10-06 10:15:15.483958207 +0000 UTC m=+5766.273983787" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.501871 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.669144 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb\") pod \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.670384 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config\") pod \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.670812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmnf4\" (UniqueName: \"kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4\") pod \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.671909 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb\") pod \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.672299 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc\") pod \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\" (UID: \"981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5\") " Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.676259 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4" (OuterVolumeSpecName: "kube-api-access-qmnf4") pod "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" (UID: "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5"). InnerVolumeSpecName "kube-api-access-qmnf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.713970 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" (UID: "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.726484 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" (UID: "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.740931 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" (UID: "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.742409 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config" (OuterVolumeSpecName: "config") pod "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" (UID: "981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.775202 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmnf4\" (UniqueName: \"kubernetes.io/projected/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-kube-api-access-qmnf4\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.775237 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.775247 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.775256 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:15 crc kubenswrapper[4989]: I1006 10:15:15.775265 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:16 crc kubenswrapper[4989]: I1006 10:15:16.434514 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-545cdf9765-8hzmr" Oct 06 10:15:16 crc kubenswrapper[4989]: I1006 10:15:16.576564 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:15:16 crc kubenswrapper[4989]: I1006 10:15:16.585783 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-545cdf9765-8hzmr"] Oct 06 10:15:16 crc kubenswrapper[4989]: I1006 10:15:16.902683 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:16.999596 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data\") pod \"3adf6a92-096f-4284-af10-9f703ff47b95\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:16.999716 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbqnf\" (UniqueName: \"kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf\") pod \"3adf6a92-096f-4284-af10-9f703ff47b95\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:16.999856 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle\") pod \"3adf6a92-096f-4284-af10-9f703ff47b95\" (UID: \"3adf6a92-096f-4284-af10-9f703ff47b95\") " Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.018905 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf" (OuterVolumeSpecName: "kube-api-access-hbqnf") pod "3adf6a92-096f-4284-af10-9f703ff47b95" (UID: "3adf6a92-096f-4284-af10-9f703ff47b95"). InnerVolumeSpecName "kube-api-access-hbqnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.043617 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data" (OuterVolumeSpecName: "config-data") pod "3adf6a92-096f-4284-af10-9f703ff47b95" (UID: "3adf6a92-096f-4284-af10-9f703ff47b95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.058422 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3adf6a92-096f-4284-af10-9f703ff47b95" (UID: "3adf6a92-096f-4284-af10-9f703ff47b95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.102795 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.103121 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3adf6a92-096f-4284-af10-9f703ff47b95-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.103213 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbqnf\" (UniqueName: \"kubernetes.io/projected/3adf6a92-096f-4284-af10-9f703ff47b95-kube-api-access-hbqnf\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.443767 4989 generic.go:334] "Generic (PLEG): container finished" podID="3adf6a92-096f-4284-af10-9f703ff47b95" containerID="30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f" exitCode=0 Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.443829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3adf6a92-096f-4284-af10-9f703ff47b95","Type":"ContainerDied","Data":"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f"} Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.444820 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3adf6a92-096f-4284-af10-9f703ff47b95","Type":"ContainerDied","Data":"fd336bafde9e0624ac0975079a501058beb3994c2e79c8e20c0ae40ebf503772"} Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.444843 4989 scope.go:117] "RemoveContainer" containerID="30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.443846 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.467867 4989 scope.go:117] "RemoveContainer" containerID="30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f" Oct 06 10:15:17 crc kubenswrapper[4989]: E1006 10:15:17.468562 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f\": container with ID starting with 30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f not found: ID does not exist" containerID="30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.468592 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f"} err="failed to get container status \"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f\": rpc error: code = NotFound desc = could not find container \"30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f\": container with ID starting with 30720cb095e72aa6879e3e64e3b116d030acf7c0f230d150e1866b2164a99f4f not found: ID does not exist" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.484938 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.495928 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.523942 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:17 crc kubenswrapper[4989]: E1006 10:15:17.524469 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="dnsmasq-dns" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.524488 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="dnsmasq-dns" Oct 06 10:15:17 crc kubenswrapper[4989]: E1006 10:15:17.524500 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adf6a92-096f-4284-af10-9f703ff47b95" containerName="nova-scheduler-scheduler" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.524506 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adf6a92-096f-4284-af10-9f703ff47b95" containerName="nova-scheduler-scheduler" Oct 06 10:15:17 crc kubenswrapper[4989]: E1006 10:15:17.524542 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="init" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.524548 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="init" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.524717 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" containerName="dnsmasq-dns" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.524737 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adf6a92-096f-4284-af10-9f703ff47b95" containerName="nova-scheduler-scheduler" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.525336 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.535593 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.540735 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.613441 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzkxb\" (UniqueName: \"kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.613610 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.613749 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.715777 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.716675 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.716823 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzkxb\" (UniqueName: \"kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.720483 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.720496 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.750178 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzkxb\" (UniqueName: \"kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb\") pod \"nova-scheduler-0\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.851103 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.951303 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3adf6a92-096f-4284-af10-9f703ff47b95" path="/var/lib/kubelet/pods/3adf6a92-096f-4284-af10-9f703ff47b95/volumes" Oct 06 10:15:17 crc kubenswrapper[4989]: I1006 10:15:17.952229 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5" path="/var/lib/kubelet/pods/981c40a8-9ddd-44dc-bd95-2e4b0fa30fe5/volumes" Oct 06 10:15:18 crc kubenswrapper[4989]: I1006 10:15:18.332380 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:18 crc kubenswrapper[4989]: I1006 10:15:18.457602 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4","Type":"ContainerStarted","Data":"29e1f767870c1782b746e5227ab402c5d32b2666c8ae5493247d10c38ce8267c"} Oct 06 10:15:18 crc kubenswrapper[4989]: I1006 10:15:18.837412 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:18 crc kubenswrapper[4989]: I1006 10:15:18.837984 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:19 crc kubenswrapper[4989]: I1006 10:15:19.472926 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4","Type":"ContainerStarted","Data":"e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2"} Oct 06 10:15:19 crc kubenswrapper[4989]: I1006 10:15:19.500696 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5006751449999998 podStartE2EDuration="2.500675145s" podCreationTimestamp="2025-10-06 10:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:19.496794463 +0000 UTC m=+5770.286820073" watchObservedRunningTime="2025-10-06 10:15:19.500675145 +0000 UTC m=+5770.290700735" Oct 06 10:15:21 crc kubenswrapper[4989]: I1006 10:15:21.026902 4989 scope.go:117] "RemoveContainer" containerID="a01613d3f1d62886f1ef7d2fbf27fd2e59e90e015b2242e39fa2ba21357d1310" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.125382 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.731791 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-hs64s"] Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.733631 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.736237 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.737144 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.746865 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-hs64s"] Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.833096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.833213 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.833353 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fsbd\" (UniqueName: \"kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.833444 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.852179 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.935276 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.936135 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.936206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fsbd\" (UniqueName: \"kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.936258 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.941396 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.941392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.942242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:22 crc kubenswrapper[4989]: I1006 10:15:22.960535 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fsbd\" (UniqueName: \"kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd\") pod \"nova-cell1-cell-mapping-hs64s\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.052101 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.499994 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-hs64s"] Oct 06 10:15:23 crc kubenswrapper[4989]: W1006 10:15:23.512867 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76629129_a886_4f22_a326_801b77cce130.slice/crio-8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6 WatchSource:0}: Error finding container 8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6: Status 404 returned error can't find the container with id 8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6 Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.534080 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-hs64s" event={"ID":"76629129-a886-4f22-a326-801b77cce130","Type":"ContainerStarted","Data":"8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6"} Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.837492 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.837887 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.863264 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:15:23 crc kubenswrapper[4989]: I1006 10:15:23.863312 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:15:24 crc kubenswrapper[4989]: I1006 10:15:24.542145 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-hs64s" event={"ID":"76629129-a886-4f22-a326-801b77cce130","Type":"ContainerStarted","Data":"8449a31a37b17ebae2efa26a1e1357e2d60608edecd6049c6e36c36856614fcd"} Oct 06 10:15:24 crc kubenswrapper[4989]: I1006 10:15:24.557805 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-hs64s" podStartSLOduration=2.557787418 podStartE2EDuration="2.557787418s" podCreationTimestamp="2025-10-06 10:15:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:24.553581327 +0000 UTC m=+5775.343606897" watchObservedRunningTime="2025-10-06 10:15:24.557787418 +0000 UTC m=+5775.347812998" Oct 06 10:15:25 crc kubenswrapper[4989]: I1006 10:15:25.003870 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.65:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:25 crc kubenswrapper[4989]: I1006 10:15:25.003906 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.66:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:25 crc kubenswrapper[4989]: I1006 10:15:25.004004 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.65:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:25 crc kubenswrapper[4989]: I1006 10:15:25.004007 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.66:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:27 crc kubenswrapper[4989]: I1006 10:15:27.852687 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:15:27 crc kubenswrapper[4989]: I1006 10:15:27.897544 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:15:28 crc kubenswrapper[4989]: I1006 10:15:28.585098 4989 generic.go:334] "Generic (PLEG): container finished" podID="76629129-a886-4f22-a326-801b77cce130" containerID="8449a31a37b17ebae2efa26a1e1357e2d60608edecd6049c6e36c36856614fcd" exitCode=0 Oct 06 10:15:28 crc kubenswrapper[4989]: I1006 10:15:28.585181 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-hs64s" event={"ID":"76629129-a886-4f22-a326-801b77cce130","Type":"ContainerDied","Data":"8449a31a37b17ebae2efa26a1e1357e2d60608edecd6049c6e36c36856614fcd"} Oct 06 10:15:28 crc kubenswrapper[4989]: I1006 10:15:28.634427 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.026813 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.100287 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fsbd\" (UniqueName: \"kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd\") pod \"76629129-a886-4f22-a326-801b77cce130\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.100363 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle\") pod \"76629129-a886-4f22-a326-801b77cce130\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.100476 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data\") pod \"76629129-a886-4f22-a326-801b77cce130\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.100631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts\") pod \"76629129-a886-4f22-a326-801b77cce130\" (UID: \"76629129-a886-4f22-a326-801b77cce130\") " Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.107405 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd" (OuterVolumeSpecName: "kube-api-access-4fsbd") pod "76629129-a886-4f22-a326-801b77cce130" (UID: "76629129-a886-4f22-a326-801b77cce130"). InnerVolumeSpecName "kube-api-access-4fsbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.122497 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts" (OuterVolumeSpecName: "scripts") pod "76629129-a886-4f22-a326-801b77cce130" (UID: "76629129-a886-4f22-a326-801b77cce130"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.129713 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data" (OuterVolumeSpecName: "config-data") pod "76629129-a886-4f22-a326-801b77cce130" (UID: "76629129-a886-4f22-a326-801b77cce130"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.137843 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76629129-a886-4f22-a326-801b77cce130" (UID: "76629129-a886-4f22-a326-801b77cce130"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.204008 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fsbd\" (UniqueName: \"kubernetes.io/projected/76629129-a886-4f22-a326-801b77cce130-kube-api-access-4fsbd\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.204039 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.204049 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.204058 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76629129-a886-4f22-a326-801b77cce130-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.652166 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-hs64s" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.652016 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-hs64s" event={"ID":"76629129-a886-4f22-a326-801b77cce130","Type":"ContainerDied","Data":"8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6"} Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.652563 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fa57d9bc305da1d32cab892542bbf7b9cb2108c54b4f5522c0cbeb6c755e6e6" Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.824262 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.824640 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-log" containerID="cri-o://608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420" gracePeriod=30 Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.824758 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-api" containerID="cri-o://233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5" gracePeriod=30 Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.840882 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.841082 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerName="nova-scheduler-scheduler" containerID="cri-o://e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" gracePeriod=30 Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.871808 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.872061 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-log" containerID="cri-o://14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b" gracePeriod=30 Oct 06 10:15:30 crc kubenswrapper[4989]: I1006 10:15:30.872551 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-metadata" containerID="cri-o://dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e" gracePeriod=30 Oct 06 10:15:31 crc kubenswrapper[4989]: I1006 10:15:31.684781 4989 generic.go:334] "Generic (PLEG): container finished" podID="e9e6e680-7d65-47bd-94af-2c930591289b" containerID="14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b" exitCode=143 Oct 06 10:15:31 crc kubenswrapper[4989]: I1006 10:15:31.684896 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerDied","Data":"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b"} Oct 06 10:15:31 crc kubenswrapper[4989]: I1006 10:15:31.687037 4989 generic.go:334] "Generic (PLEG): container finished" podID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerID="608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420" exitCode=143 Oct 06 10:15:31 crc kubenswrapper[4989]: I1006 10:15:31.687086 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerDied","Data":"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420"} Oct 06 10:15:32 crc kubenswrapper[4989]: E1006 10:15:32.854807 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:15:32 crc kubenswrapper[4989]: E1006 10:15:32.857024 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:15:32 crc kubenswrapper[4989]: E1006 10:15:32.858846 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 10:15:32 crc kubenswrapper[4989]: E1006 10:15:32.858884 4989 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerName="nova-scheduler-scheduler" Oct 06 10:15:33 crc kubenswrapper[4989]: I1006 10:15:33.935849 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:15:33 crc kubenswrapper[4989]: I1006 10:15:33.935923 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.492164 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.508019 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.595421 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdcr2\" (UniqueName: \"kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2\") pod \"e9e6e680-7d65-47bd-94af-2c930591289b\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.595504 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hcg2\" (UniqueName: \"kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2\") pod \"fb1065a0-9d18-4c3f-af58-8a1a99948144\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.595533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle\") pod \"fb1065a0-9d18-4c3f-af58-8a1a99948144\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.596220 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle\") pod \"e9e6e680-7d65-47bd-94af-2c930591289b\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.596340 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data\") pod \"e9e6e680-7d65-47bd-94af-2c930591289b\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.596615 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data\") pod \"fb1065a0-9d18-4c3f-af58-8a1a99948144\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.596647 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs\") pod \"fb1065a0-9d18-4c3f-af58-8a1a99948144\" (UID: \"fb1065a0-9d18-4c3f-af58-8a1a99948144\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.596766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs\") pod \"e9e6e680-7d65-47bd-94af-2c930591289b\" (UID: \"e9e6e680-7d65-47bd-94af-2c930591289b\") " Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.597166 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs" (OuterVolumeSpecName: "logs") pod "fb1065a0-9d18-4c3f-af58-8a1a99948144" (UID: "fb1065a0-9d18-4c3f-af58-8a1a99948144"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.597735 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs" (OuterVolumeSpecName: "logs") pod "e9e6e680-7d65-47bd-94af-2c930591289b" (UID: "e9e6e680-7d65-47bd-94af-2c930591289b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.600918 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2" (OuterVolumeSpecName: "kube-api-access-gdcr2") pod "e9e6e680-7d65-47bd-94af-2c930591289b" (UID: "e9e6e680-7d65-47bd-94af-2c930591289b"). InnerVolumeSpecName "kube-api-access-gdcr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.601350 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2" (OuterVolumeSpecName: "kube-api-access-4hcg2") pod "fb1065a0-9d18-4c3f-af58-8a1a99948144" (UID: "fb1065a0-9d18-4c3f-af58-8a1a99948144"). InnerVolumeSpecName "kube-api-access-4hcg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.621073 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data" (OuterVolumeSpecName: "config-data") pod "fb1065a0-9d18-4c3f-af58-8a1a99948144" (UID: "fb1065a0-9d18-4c3f-af58-8a1a99948144"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.621815 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data" (OuterVolumeSpecName: "config-data") pod "e9e6e680-7d65-47bd-94af-2c930591289b" (UID: "e9e6e680-7d65-47bd-94af-2c930591289b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.623218 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb1065a0-9d18-4c3f-af58-8a1a99948144" (UID: "fb1065a0-9d18-4c3f-af58-8a1a99948144"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.661140 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9e6e680-7d65-47bd-94af-2c930591289b" (UID: "e9e6e680-7d65-47bd-94af-2c930591289b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698678 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698728 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb1065a0-9d18-4c3f-af58-8a1a99948144-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698738 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9e6e680-7d65-47bd-94af-2c930591289b-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698747 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdcr2\" (UniqueName: \"kubernetes.io/projected/e9e6e680-7d65-47bd-94af-2c930591289b-kube-api-access-gdcr2\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698757 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hcg2\" (UniqueName: \"kubernetes.io/projected/fb1065a0-9d18-4c3f-af58-8a1a99948144-kube-api-access-4hcg2\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698765 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb1065a0-9d18-4c3f-af58-8a1a99948144-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698774 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.698783 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9e6e680-7d65-47bd-94af-2c930591289b-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.722524 4989 generic.go:334] "Generic (PLEG): container finished" podID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerID="e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" exitCode=0 Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.722555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4","Type":"ContainerDied","Data":"e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2"} Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.724833 4989 generic.go:334] "Generic (PLEG): container finished" podID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerID="233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5" exitCode=0 Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.725043 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerDied","Data":"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5"} Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.725178 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fb1065a0-9d18-4c3f-af58-8a1a99948144","Type":"ContainerDied","Data":"c5fab9b63d52a8a9fb29d5975a5dbb1b14d8d1ab78ca96fb9ed74703652fdb0d"} Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.725300 4989 scope.go:117] "RemoveContainer" containerID="233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.725587 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.733867 4989 generic.go:334] "Generic (PLEG): container finished" podID="e9e6e680-7d65-47bd-94af-2c930591289b" containerID="dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e" exitCode=0 Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.733905 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerDied","Data":"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e"} Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.733912 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.733929 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9e6e680-7d65-47bd-94af-2c930591289b","Type":"ContainerDied","Data":"d8bd3410e22a5fbd594a3142a0fa3deaff61e32f70a4dd0a776f1300a15e85c2"} Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.757769 4989 scope.go:117] "RemoveContainer" containerID="608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.783517 4989 scope.go:117] "RemoveContainer" containerID="233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.784006 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5\": container with ID starting with 233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5 not found: ID does not exist" containerID="233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.784039 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5"} err="failed to get container status \"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5\": rpc error: code = NotFound desc = could not find container \"233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5\": container with ID starting with 233c17ee536746e4e28cba6a2fabad925cbd03d102b421eaa85a06782f677bb5 not found: ID does not exist" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.784063 4989 scope.go:117] "RemoveContainer" containerID="608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.784639 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420\": container with ID starting with 608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420 not found: ID does not exist" containerID="608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.784669 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420"} err="failed to get container status \"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420\": rpc error: code = NotFound desc = could not find container \"608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420\": container with ID starting with 608b2be7bea720182b6bcf2b529a63d59ccd90ffc3fecf66e53948a98a8e9420 not found: ID does not exist" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.784703 4989 scope.go:117] "RemoveContainer" containerID="dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.785412 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.812539 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.833169 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.843189 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.846419 4989 scope.go:117] "RemoveContainer" containerID="14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.852447 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.852904 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-log" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.852919 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-log" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.852944 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76629129-a886-4f22-a326-801b77cce130" containerName="nova-manage" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.852953 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="76629129-a886-4f22-a326-801b77cce130" containerName="nova-manage" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.852967 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-metadata" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.852976 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-metadata" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.852998 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-log" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853007 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-log" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.853019 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-api" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853027 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-api" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853226 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-log" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853240 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-metadata" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853267 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" containerName="nova-metadata-log" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853279 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" containerName="nova-api-api" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.853298 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="76629129-a886-4f22-a326-801b77cce130" containerName="nova-manage" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.854404 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.856626 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.862675 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.864267 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.867901 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.878179 4989 scope.go:117] "RemoveContainer" containerID="dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.881086 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e\": container with ID starting with dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e not found: ID does not exist" containerID="dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.881222 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e"} err="failed to get container status \"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e\": rpc error: code = NotFound desc = could not find container \"dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e\": container with ID starting with dfac718c087d2cddecedcdc4a931325b5989754961b68fbcebad05b744b8467e not found: ID does not exist" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.881262 4989 scope.go:117] "RemoveContainer" containerID="14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b" Oct 06 10:15:34 crc kubenswrapper[4989]: E1006 10:15:34.882559 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b\": container with ID starting with 14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b not found: ID does not exist" containerID="14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.882647 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b"} err="failed to get container status \"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b\": rpc error: code = NotFound desc = could not find container \"14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b\": container with ID starting with 14fbc69e415ccd42e786268de8a768f8239620bbb2036ec91dc44df84309747b not found: ID does not exist" Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.892108 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.928106 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:34 crc kubenswrapper[4989]: I1006 10:15:34.977905 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.004983 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005029 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005059 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005106 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gppzg\" (UniqueName: \"kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005185 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005214 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005264 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xmcz\" (UniqueName: \"kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.005306 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.106578 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data\") pod \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.106924 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzkxb\" (UniqueName: \"kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb\") pod \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107038 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle\") pod \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\" (UID: \"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4\") " Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107337 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107453 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107479 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gppzg\" (UniqueName: \"kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107617 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.107664 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xmcz\" (UniqueName: \"kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.108693 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.110915 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.117647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.118077 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb" (OuterVolumeSpecName: "kube-api-access-bzkxb") pod "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" (UID: "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4"). InnerVolumeSpecName "kube-api-access-bzkxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.118490 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.119906 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.127862 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xmcz\" (UniqueName: \"kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.131729 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data\") pod \"nova-metadata-0\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.135116 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gppzg\" (UniqueName: \"kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg\") pod \"nova-api-0\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.139676 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data" (OuterVolumeSpecName: "config-data") pod "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" (UID: "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.141522 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" (UID: "df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.192579 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.207552 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.209154 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.209185 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzkxb\" (UniqueName: \"kubernetes.io/projected/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-kube-api-access-bzkxb\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.209198 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:15:35 crc kubenswrapper[4989]: W1006 10:15:35.691477 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ab7d5dd_7194_48f0_826d_4c0e986a28a6.slice/crio-404626d57d95245f3818a1d98ac9864022ed0f5738e489b1122af399ae7b4ab9 WatchSource:0}: Error finding container 404626d57d95245f3818a1d98ac9864022ed0f5738e489b1122af399ae7b4ab9: Status 404 returned error can't find the container with id 404626d57d95245f3818a1d98ac9864022ed0f5738e489b1122af399ae7b4ab9 Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.692080 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.756617 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.758438 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4","Type":"ContainerDied","Data":"29e1f767870c1782b746e5227ab402c5d32b2666c8ae5493247d10c38ce8267c"} Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.758502 4989 scope.go:117] "RemoveContainer" containerID="e6b6c61d9ed297aee2f815f43f023c19080c489eca24efcfaf640b087552f9b2" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.758498 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.761037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerStarted","Data":"404626d57d95245f3818a1d98ac9864022ed0f5738e489b1122af399ae7b4ab9"} Oct 06 10:15:35 crc kubenswrapper[4989]: W1006 10:15:35.773054 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14a26d46_0479_4e10_92c9_2479a97b5218.slice/crio-d1a166308e0bc354ea46b758d16497d1187eb20de1b3ec2949cc95043b8f9023 WatchSource:0}: Error finding container d1a166308e0bc354ea46b758d16497d1187eb20de1b3ec2949cc95043b8f9023: Status 404 returned error can't find the container with id d1a166308e0bc354ea46b758d16497d1187eb20de1b3ec2949cc95043b8f9023 Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.797703 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.810491 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.825360 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: E1006 10:15:35.825757 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerName="nova-scheduler-scheduler" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.825774 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerName="nova-scheduler-scheduler" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.825955 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" containerName="nova-scheduler-scheduler" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.826550 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.830458 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.835696 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.923717 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s6rw\" (UniqueName: \"kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.923774 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.923943 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.948714 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4" path="/var/lib/kubelet/pods/df6f8e72-1d07-4cc0-aa48-dea3b9e4aae4/volumes" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.949812 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9e6e680-7d65-47bd-94af-2c930591289b" path="/var/lib/kubelet/pods/e9e6e680-7d65-47bd-94af-2c930591289b/volumes" Oct 06 10:15:35 crc kubenswrapper[4989]: I1006 10:15:35.950544 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb1065a0-9d18-4c3f-af58-8a1a99948144" path="/var/lib/kubelet/pods/fb1065a0-9d18-4c3f-af58-8a1a99948144/volumes" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.025268 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.025337 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s6rw\" (UniqueName: \"kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.025374 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.030458 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.032191 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.048668 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s6rw\" (UniqueName: \"kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw\") pod \"nova-scheduler-0\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.158595 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.642196 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:15:36 crc kubenswrapper[4989]: W1006 10:15:36.643514 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50c10569_bfee_46ef_bdf0_bcd8aea02659.slice/crio-7e3801f61d4fc90fb1938f06abd10f882a1713b5fdb9644ffb070d8ea098bae4 WatchSource:0}: Error finding container 7e3801f61d4fc90fb1938f06abd10f882a1713b5fdb9644ffb070d8ea098bae4: Status 404 returned error can't find the container with id 7e3801f61d4fc90fb1938f06abd10f882a1713b5fdb9644ffb070d8ea098bae4 Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.775629 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50c10569-bfee-46ef-bdf0-bcd8aea02659","Type":"ContainerStarted","Data":"7e3801f61d4fc90fb1938f06abd10f882a1713b5fdb9644ffb070d8ea098bae4"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.780805 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerStarted","Data":"4d03966f08e652c143bac94ca7afe54a990d2ebe508f8d91c76de6cd995bd4d7"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.780864 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerStarted","Data":"7e2bba5ca8017406b6db629faa03406fef22830bae3ed86b9eb168efad28cf5a"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.785151 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerStarted","Data":"11f723384ee2b989d4b8b4e11647be6158de5fa6075da0134eda587a59a800c3"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.785198 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerStarted","Data":"daf18c273de7a1b7a87ec84169ae813b9b36f24ad44f46542fc90923855737db"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.785217 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerStarted","Data":"d1a166308e0bc354ea46b758d16497d1187eb20de1b3ec2949cc95043b8f9023"} Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.812011 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.811992343 podStartE2EDuration="2.811992343s" podCreationTimestamp="2025-10-06 10:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:36.803154029 +0000 UTC m=+5787.593179619" watchObservedRunningTime="2025-10-06 10:15:36.811992343 +0000 UTC m=+5787.602017933" Oct 06 10:15:36 crc kubenswrapper[4989]: I1006 10:15:36.846435 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.846406694 podStartE2EDuration="2.846406694s" podCreationTimestamp="2025-10-06 10:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:36.845258351 +0000 UTC m=+5787.635283951" watchObservedRunningTime="2025-10-06 10:15:36.846406694 +0000 UTC m=+5787.636432284" Oct 06 10:15:37 crc kubenswrapper[4989]: I1006 10:15:37.801811 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50c10569-bfee-46ef-bdf0-bcd8aea02659","Type":"ContainerStarted","Data":"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2"} Oct 06 10:15:37 crc kubenswrapper[4989]: I1006 10:15:37.837983 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8379581849999997 podStartE2EDuration="2.837958185s" podCreationTimestamp="2025-10-06 10:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:37.827544115 +0000 UTC m=+5788.617569735" watchObservedRunningTime="2025-10-06 10:15:37.837958185 +0000 UTC m=+5788.627983795" Oct 06 10:15:40 crc kubenswrapper[4989]: I1006 10:15:40.208506 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:40 crc kubenswrapper[4989]: I1006 10:15:40.208930 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:15:41 crc kubenswrapper[4989]: I1006 10:15:41.159760 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:15:45 crc kubenswrapper[4989]: I1006 10:15:45.193947 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:15:45 crc kubenswrapper[4989]: I1006 10:15:45.194548 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:15:45 crc kubenswrapper[4989]: I1006 10:15:45.208070 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:15:45 crc kubenswrapper[4989]: I1006 10:15:45.208499 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.159680 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.193274 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.359884 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.359940 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.69:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.360041 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.359903 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:15:46 crc kubenswrapper[4989]: I1006 10:15:46.942897 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.197394 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.198709 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.201928 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.203591 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.211186 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.212010 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:15:55 crc kubenswrapper[4989]: I1006 10:15:55.214092 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.024013 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.026225 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.028841 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.273615 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.276620 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.289712 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.356824 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.356891 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.357002 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.357031 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.357069 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggz7m\" (UniqueName: \"kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.458624 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.458718 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.458815 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.458841 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.458880 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggz7m\" (UniqueName: \"kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.459520 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.459712 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.459719 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.459822 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.478641 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggz7m\" (UniqueName: \"kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m\") pod \"dnsmasq-dns-845cdfbdfc-8jssc\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.602186 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:56 crc kubenswrapper[4989]: I1006 10:15:56.862701 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:15:57 crc kubenswrapper[4989]: I1006 10:15:57.034863 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" event={"ID":"1ff41850-0386-4b7e-98b1-2fff72b7b784","Type":"ContainerStarted","Data":"b19302938f5d02eb47a8424e34275436a59f81eec715cef255662a028f778641"} Oct 06 10:15:58 crc kubenswrapper[4989]: I1006 10:15:58.044484 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerID="a6c978abbacda0a19a12c002a7cc830f2778e9490bebc33a80f2e691fe8e4656" exitCode=0 Oct 06 10:15:58 crc kubenswrapper[4989]: I1006 10:15:58.044533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" event={"ID":"1ff41850-0386-4b7e-98b1-2fff72b7b784","Type":"ContainerDied","Data":"a6c978abbacda0a19a12c002a7cc830f2778e9490bebc33a80f2e691fe8e4656"} Oct 06 10:15:59 crc kubenswrapper[4989]: I1006 10:15:59.057789 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" event={"ID":"1ff41850-0386-4b7e-98b1-2fff72b7b784","Type":"ContainerStarted","Data":"d03d5347d142dfcac09d25cd861f12115a13e1578f1c3c9b7df9007ff48ed6b7"} Oct 06 10:15:59 crc kubenswrapper[4989]: I1006 10:15:59.057862 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:15:59 crc kubenswrapper[4989]: I1006 10:15:59.091977 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" podStartSLOduration=3.091949961 podStartE2EDuration="3.091949961s" podCreationTimestamp="2025-10-06 10:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:15:59.080972045 +0000 UTC m=+5809.870997645" watchObservedRunningTime="2025-10-06 10:15:59.091949961 +0000 UTC m=+5809.881975551" Oct 06 10:16:03 crc kubenswrapper[4989]: I1006 10:16:03.935600 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:16:03 crc kubenswrapper[4989]: I1006 10:16:03.936249 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:16:03 crc kubenswrapper[4989]: I1006 10:16:03.936318 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:16:03 crc kubenswrapper[4989]: I1006 10:16:03.938128 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:16:03 crc kubenswrapper[4989]: I1006 10:16:03.938198 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0" gracePeriod=600 Oct 06 10:16:04 crc kubenswrapper[4989]: I1006 10:16:04.122299 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0" exitCode=0 Oct 06 10:16:04 crc kubenswrapper[4989]: I1006 10:16:04.122382 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0"} Oct 06 10:16:04 crc kubenswrapper[4989]: I1006 10:16:04.122612 4989 scope.go:117] "RemoveContainer" containerID="2a88a83ff649a5a676e745d61641d25920f8956b01c5139e91689eec2b22e5d5" Oct 06 10:16:05 crc kubenswrapper[4989]: I1006 10:16:05.132669 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb"} Oct 06 10:16:06 crc kubenswrapper[4989]: I1006 10:16:06.604981 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:16:06 crc kubenswrapper[4989]: I1006 10:16:06.702379 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:16:06 crc kubenswrapper[4989]: I1006 10:16:06.702686 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="dnsmasq-dns" containerID="cri-o://4b7850684d23fdd264dd1b306607c85ba0abcb31b67d45fec01245852f5b575b" gracePeriod=10 Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.155072 4989 generic.go:334] "Generic (PLEG): container finished" podID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerID="4b7850684d23fdd264dd1b306607c85ba0abcb31b67d45fec01245852f5b575b" exitCode=0 Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.155146 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" event={"ID":"9cce2f3f-e874-4c9d-805f-784219fd05d5","Type":"ContainerDied","Data":"4b7850684d23fdd264dd1b306607c85ba0abcb31b67d45fec01245852f5b575b"} Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.155348 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" event={"ID":"9cce2f3f-e874-4c9d-805f-784219fd05d5","Type":"ContainerDied","Data":"b915c0cf5a3d1a7e419845f23011a9d60119b7eb81d8b704f9f2cca5be569bd7"} Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.155363 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b915c0cf5a3d1a7e419845f23011a9d60119b7eb81d8b704f9f2cca5be569bd7" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.229745 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.273073 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfgh7\" (UniqueName: \"kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7\") pod \"9cce2f3f-e874-4c9d-805f-784219fd05d5\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.273137 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc\") pod \"9cce2f3f-e874-4c9d-805f-784219fd05d5\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.273229 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb\") pod \"9cce2f3f-e874-4c9d-805f-784219fd05d5\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.273337 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb\") pod \"9cce2f3f-e874-4c9d-805f-784219fd05d5\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.273444 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config\") pod \"9cce2f3f-e874-4c9d-805f-784219fd05d5\" (UID: \"9cce2f3f-e874-4c9d-805f-784219fd05d5\") " Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.284968 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7" (OuterVolumeSpecName: "kube-api-access-sfgh7") pod "9cce2f3f-e874-4c9d-805f-784219fd05d5" (UID: "9cce2f3f-e874-4c9d-805f-784219fd05d5"). InnerVolumeSpecName "kube-api-access-sfgh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.336337 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9cce2f3f-e874-4c9d-805f-784219fd05d5" (UID: "9cce2f3f-e874-4c9d-805f-784219fd05d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.342063 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9cce2f3f-e874-4c9d-805f-784219fd05d5" (UID: "9cce2f3f-e874-4c9d-805f-784219fd05d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.348859 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config" (OuterVolumeSpecName: "config") pod "9cce2f3f-e874-4c9d-805f-784219fd05d5" (UID: "9cce2f3f-e874-4c9d-805f-784219fd05d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.361574 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9cce2f3f-e874-4c9d-805f-784219fd05d5" (UID: "9cce2f3f-e874-4c9d-805f-784219fd05d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.376572 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.376648 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfgh7\" (UniqueName: \"kubernetes.io/projected/9cce2f3f-e874-4c9d-805f-784219fd05d5-kube-api-access-sfgh7\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.376710 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.376728 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:07 crc kubenswrapper[4989]: I1006 10:16:07.376780 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9cce2f3f-e874-4c9d-805f-784219fd05d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:08 crc kubenswrapper[4989]: I1006 10:16:08.165887 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf6b7bb9-h75b4" Oct 06 10:16:08 crc kubenswrapper[4989]: I1006 10:16:08.199705 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:16:08 crc kubenswrapper[4989]: I1006 10:16:08.207062 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dcf6b7bb9-h75b4"] Oct 06 10:16:09 crc kubenswrapper[4989]: I1006 10:16:09.950836 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" path="/var/lib/kubelet/pods/9cce2f3f-e874-4c9d-805f-784219fd05d5/volumes" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.116021 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bl89j"] Oct 06 10:16:10 crc kubenswrapper[4989]: E1006 10:16:10.116673 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="dnsmasq-dns" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.116738 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="dnsmasq-dns" Oct 06 10:16:10 crc kubenswrapper[4989]: E1006 10:16:10.116817 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="init" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.116866 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="init" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.117070 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cce2f3f-e874-4c9d-805f-784219fd05d5" containerName="dnsmasq-dns" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.117852 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.149106 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bl89j"] Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.245939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k724\" (UniqueName: \"kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724\") pod \"cinder-db-create-bl89j\" (UID: \"d65adde8-6511-4eae-8bc5-0ca4980761fb\") " pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.348483 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k724\" (UniqueName: \"kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724\") pod \"cinder-db-create-bl89j\" (UID: \"d65adde8-6511-4eae-8bc5-0ca4980761fb\") " pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.370504 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k724\" (UniqueName: \"kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724\") pod \"cinder-db-create-bl89j\" (UID: \"d65adde8-6511-4eae-8bc5-0ca4980761fb\") " pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.452307 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:10 crc kubenswrapper[4989]: I1006 10:16:10.937696 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bl89j"] Oct 06 10:16:10 crc kubenswrapper[4989]: W1006 10:16:10.943306 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd65adde8_6511_4eae_8bc5_0ca4980761fb.slice/crio-d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09 WatchSource:0}: Error finding container d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09: Status 404 returned error can't find the container with id d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09 Oct 06 10:16:11 crc kubenswrapper[4989]: I1006 10:16:11.198378 4989 generic.go:334] "Generic (PLEG): container finished" podID="d65adde8-6511-4eae-8bc5-0ca4980761fb" containerID="85733463780b84027e139388a1aa6544e9a1e472a5326431061b545fed7e66a6" exitCode=0 Oct 06 10:16:11 crc kubenswrapper[4989]: I1006 10:16:11.198427 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bl89j" event={"ID":"d65adde8-6511-4eae-8bc5-0ca4980761fb","Type":"ContainerDied","Data":"85733463780b84027e139388a1aa6544e9a1e472a5326431061b545fed7e66a6"} Oct 06 10:16:11 crc kubenswrapper[4989]: I1006 10:16:11.198778 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bl89j" event={"ID":"d65adde8-6511-4eae-8bc5-0ca4980761fb","Type":"ContainerStarted","Data":"d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09"} Oct 06 10:16:12 crc kubenswrapper[4989]: I1006 10:16:12.636603 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:12 crc kubenswrapper[4989]: I1006 10:16:12.797202 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k724\" (UniqueName: \"kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724\") pod \"d65adde8-6511-4eae-8bc5-0ca4980761fb\" (UID: \"d65adde8-6511-4eae-8bc5-0ca4980761fb\") " Oct 06 10:16:12 crc kubenswrapper[4989]: I1006 10:16:12.805554 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724" (OuterVolumeSpecName: "kube-api-access-5k724") pod "d65adde8-6511-4eae-8bc5-0ca4980761fb" (UID: "d65adde8-6511-4eae-8bc5-0ca4980761fb"). InnerVolumeSpecName "kube-api-access-5k724". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:12 crc kubenswrapper[4989]: I1006 10:16:12.900161 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k724\" (UniqueName: \"kubernetes.io/projected/d65adde8-6511-4eae-8bc5-0ca4980761fb-kube-api-access-5k724\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:13 crc kubenswrapper[4989]: I1006 10:16:13.221472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bl89j" event={"ID":"d65adde8-6511-4eae-8bc5-0ca4980761fb","Type":"ContainerDied","Data":"d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09"} Oct 06 10:16:13 crc kubenswrapper[4989]: I1006 10:16:13.221515 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9bf6beed6fbcd5b886c0bb0bac979aee677b5c62ad69d5e4c93f2f2f8fa3a09" Oct 06 10:16:13 crc kubenswrapper[4989]: I1006 10:16:13.221532 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bl89j" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.223539 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-079f-account-create-jj5g2"] Oct 06 10:16:20 crc kubenswrapper[4989]: E1006 10:16:20.224849 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65adde8-6511-4eae-8bc5-0ca4980761fb" containerName="mariadb-database-create" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.224877 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65adde8-6511-4eae-8bc5-0ca4980761fb" containerName="mariadb-database-create" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.225239 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65adde8-6511-4eae-8bc5-0ca4980761fb" containerName="mariadb-database-create" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.226283 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.229325 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.242223 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-079f-account-create-jj5g2"] Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.350392 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhqqd\" (UniqueName: \"kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd\") pod \"cinder-079f-account-create-jj5g2\" (UID: \"e46a0615-6b06-48e6-9769-43533448ae24\") " pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.452206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhqqd\" (UniqueName: \"kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd\") pod \"cinder-079f-account-create-jj5g2\" (UID: \"e46a0615-6b06-48e6-9769-43533448ae24\") " pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.478051 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhqqd\" (UniqueName: \"kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd\") pod \"cinder-079f-account-create-jj5g2\" (UID: \"e46a0615-6b06-48e6-9769-43533448ae24\") " pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:20 crc kubenswrapper[4989]: I1006 10:16:20.557497 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:21 crc kubenswrapper[4989]: I1006 10:16:21.057950 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-079f-account-create-jj5g2"] Oct 06 10:16:21 crc kubenswrapper[4989]: I1006 10:16:21.309521 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-079f-account-create-jj5g2" event={"ID":"e46a0615-6b06-48e6-9769-43533448ae24","Type":"ContainerStarted","Data":"809031e81d6d9c206c8ec14d534123f7be95923b5c67f266ec8f65b230e05359"} Oct 06 10:16:22 crc kubenswrapper[4989]: I1006 10:16:22.323394 4989 generic.go:334] "Generic (PLEG): container finished" podID="e46a0615-6b06-48e6-9769-43533448ae24" containerID="6191a08b492a84f18b8fc40aba656a9636d305478e190fa06977403750679adc" exitCode=0 Oct 06 10:16:22 crc kubenswrapper[4989]: I1006 10:16:22.323547 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-079f-account-create-jj5g2" event={"ID":"e46a0615-6b06-48e6-9769-43533448ae24","Type":"ContainerDied","Data":"6191a08b492a84f18b8fc40aba656a9636d305478e190fa06977403750679adc"} Oct 06 10:16:23 crc kubenswrapper[4989]: I1006 10:16:23.790674 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:23 crc kubenswrapper[4989]: I1006 10:16:23.921523 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhqqd\" (UniqueName: \"kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd\") pod \"e46a0615-6b06-48e6-9769-43533448ae24\" (UID: \"e46a0615-6b06-48e6-9769-43533448ae24\") " Oct 06 10:16:23 crc kubenswrapper[4989]: I1006 10:16:23.931019 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd" (OuterVolumeSpecName: "kube-api-access-xhqqd") pod "e46a0615-6b06-48e6-9769-43533448ae24" (UID: "e46a0615-6b06-48e6-9769-43533448ae24"). InnerVolumeSpecName "kube-api-access-xhqqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:24 crc kubenswrapper[4989]: I1006 10:16:24.023536 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhqqd\" (UniqueName: \"kubernetes.io/projected/e46a0615-6b06-48e6-9769-43533448ae24-kube-api-access-xhqqd\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:24 crc kubenswrapper[4989]: I1006 10:16:24.353019 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-079f-account-create-jj5g2" event={"ID":"e46a0615-6b06-48e6-9769-43533448ae24","Type":"ContainerDied","Data":"809031e81d6d9c206c8ec14d534123f7be95923b5c67f266ec8f65b230e05359"} Oct 06 10:16:24 crc kubenswrapper[4989]: I1006 10:16:24.353095 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="809031e81d6d9c206c8ec14d534123f7be95923b5c67f266ec8f65b230e05359" Oct 06 10:16:24 crc kubenswrapper[4989]: I1006 10:16:24.353198 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-079f-account-create-jj5g2" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.489568 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-j76hm"] Oct 06 10:16:25 crc kubenswrapper[4989]: E1006 10:16:25.490496 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46a0615-6b06-48e6-9769-43533448ae24" containerName="mariadb-account-create" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.490519 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46a0615-6b06-48e6-9769-43533448ae24" containerName="mariadb-account-create" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.490880 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e46a0615-6b06-48e6-9769-43533448ae24" containerName="mariadb-account-create" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.492171 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.495850 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.496190 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.500501 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7qrzh" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.507467 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-j76hm"] Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560460 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560541 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560590 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt2ht\" (UniqueName: \"kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560828 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.560873 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.663562 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.663671 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.663724 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.663826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.664311 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.664521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.664877 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt2ht\" (UniqueName: \"kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.671141 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.671386 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.671742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.676324 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.692450 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt2ht\" (UniqueName: \"kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht\") pod \"cinder-db-sync-j76hm\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:25 crc kubenswrapper[4989]: I1006 10:16:25.817769 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:26 crc kubenswrapper[4989]: I1006 10:16:26.326013 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-j76hm"] Oct 06 10:16:26 crc kubenswrapper[4989]: I1006 10:16:26.374878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-j76hm" event={"ID":"e276c28a-7ef5-4a0d-ae40-11603f6a647e","Type":"ContainerStarted","Data":"c94472554e3d362bcc191c3039651743833fec29a5de9b3fcd2a247404b70253"} Oct 06 10:16:27 crc kubenswrapper[4989]: I1006 10:16:27.387545 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-j76hm" event={"ID":"e276c28a-7ef5-4a0d-ae40-11603f6a647e","Type":"ContainerStarted","Data":"ef12b1feead78763aac37daf3045e69df1156a52ac43b34dad9df053685b0003"} Oct 06 10:16:27 crc kubenswrapper[4989]: I1006 10:16:27.410065 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-j76hm" podStartSLOduration=2.410039981 podStartE2EDuration="2.410039981s" podCreationTimestamp="2025-10-06 10:16:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:27.406486678 +0000 UTC m=+5838.196512298" watchObservedRunningTime="2025-10-06 10:16:27.410039981 +0000 UTC m=+5838.200065601" Oct 06 10:16:29 crc kubenswrapper[4989]: I1006 10:16:29.415457 4989 generic.go:334] "Generic (PLEG): container finished" podID="e276c28a-7ef5-4a0d-ae40-11603f6a647e" containerID="ef12b1feead78763aac37daf3045e69df1156a52ac43b34dad9df053685b0003" exitCode=0 Oct 06 10:16:29 crc kubenswrapper[4989]: I1006 10:16:29.415589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-j76hm" event={"ID":"e276c28a-7ef5-4a0d-ae40-11603f6a647e","Type":"ContainerDied","Data":"ef12b1feead78763aac37daf3045e69df1156a52ac43b34dad9df053685b0003"} Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.814324 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.892811 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.893224 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.893294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.893387 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt2ht\" (UniqueName: \"kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.893413 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.893485 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id\") pod \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\" (UID: \"e276c28a-7ef5-4a0d-ae40-11603f6a647e\") " Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.894314 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.903996 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.904636 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht" (OuterVolumeSpecName: "kube-api-access-rt2ht") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "kube-api-access-rt2ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.922339 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts" (OuterVolumeSpecName: "scripts") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.964240 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data" (OuterVolumeSpecName: "config-data") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.973083 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e276c28a-7ef5-4a0d-ae40-11603f6a647e" (UID: "e276c28a-7ef5-4a0d-ae40-11603f6a647e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996373 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996409 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996421 4989 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996433 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt2ht\" (UniqueName: \"kubernetes.io/projected/e276c28a-7ef5-4a0d-ae40-11603f6a647e-kube-api-access-rt2ht\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996444 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e276c28a-7ef5-4a0d-ae40-11603f6a647e-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:30 crc kubenswrapper[4989]: I1006 10:16:30.996455 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e276c28a-7ef5-4a0d-ae40-11603f6a647e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.443807 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-j76hm" event={"ID":"e276c28a-7ef5-4a0d-ae40-11603f6a647e","Type":"ContainerDied","Data":"c94472554e3d362bcc191c3039651743833fec29a5de9b3fcd2a247404b70253"} Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.443868 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c94472554e3d362bcc191c3039651743833fec29a5de9b3fcd2a247404b70253" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.443870 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-j76hm" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.759465 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:16:31 crc kubenswrapper[4989]: E1006 10:16:31.759955 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e276c28a-7ef5-4a0d-ae40-11603f6a647e" containerName="cinder-db-sync" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.759973 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e276c28a-7ef5-4a0d-ae40-11603f6a647e" containerName="cinder-db-sync" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.760889 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e276c28a-7ef5-4a0d-ae40-11603f6a647e" containerName="cinder-db-sync" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.761986 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.776849 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.879957 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.882198 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.894858 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7qrzh" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.895645 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.895848 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.896115 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.897968 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.918178 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2mwl\" (UniqueName: \"kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.918220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.918277 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.918326 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:31 crc kubenswrapper[4989]: I1006 10:16:31.918348 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020059 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020388 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020421 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020502 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020562 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhzwq\" (UniqueName: \"kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020594 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020647 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2mwl\" (UniqueName: \"kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020697 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020750 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.020790 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.021492 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.021718 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.021867 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.022358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.038396 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2mwl\" (UniqueName: \"kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl\") pod \"dnsmasq-dns-5c8cff6cc7-h6j8v\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.102907 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.121976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122043 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122093 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122113 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122169 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhzwq\" (UniqueName: \"kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122192 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122807 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.122817 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.127358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.128102 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.128323 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.128586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.145059 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhzwq\" (UniqueName: \"kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq\") pod \"cinder-api-0\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.212850 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.599588 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:16:32 crc kubenswrapper[4989]: I1006 10:16:32.732938 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:16:32 crc kubenswrapper[4989]: W1006 10:16:32.739351 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod264b1db1_3250_43e1_9928_04cd00bb5052.slice/crio-2762ec505277cdd98dcd3f14f068e3db71b1024af2589622a9650d54e561fb89 WatchSource:0}: Error finding container 2762ec505277cdd98dcd3f14f068e3db71b1024af2589622a9650d54e561fb89: Status 404 returned error can't find the container with id 2762ec505277cdd98dcd3f14f068e3db71b1024af2589622a9650d54e561fb89 Oct 06 10:16:33 crc kubenswrapper[4989]: I1006 10:16:33.471854 4989 generic.go:334] "Generic (PLEG): container finished" podID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerID="21fdedabdced0c757245207ded0d806336f2b5cdf44bdd87a691f5f7419c4637" exitCode=0 Oct 06 10:16:33 crc kubenswrapper[4989]: I1006 10:16:33.472001 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" event={"ID":"fb0ff807-5082-4bde-b6c7-f359964df63a","Type":"ContainerDied","Data":"21fdedabdced0c757245207ded0d806336f2b5cdf44bdd87a691f5f7419c4637"} Oct 06 10:16:33 crc kubenswrapper[4989]: I1006 10:16:33.472134 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" event={"ID":"fb0ff807-5082-4bde-b6c7-f359964df63a","Type":"ContainerStarted","Data":"7605644d43c475bcedc324d0ea31f2c7235428a894d7f1e2301132c4d36e8a50"} Oct 06 10:16:33 crc kubenswrapper[4989]: I1006 10:16:33.474080 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerStarted","Data":"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235"} Oct 06 10:16:33 crc kubenswrapper[4989]: I1006 10:16:33.474110 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerStarted","Data":"2762ec505277cdd98dcd3f14f068e3db71b1024af2589622a9650d54e561fb89"} Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.503131 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerStarted","Data":"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6"} Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.506675 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.522239 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" event={"ID":"fb0ff807-5082-4bde-b6c7-f359964df63a","Type":"ContainerStarted","Data":"f4cabf623c76ab6568f152e7040eac6c77fdc1d74f352e719e51039b404d1d2b"} Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.522548 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.573895 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" podStartSLOduration=3.573868644 podStartE2EDuration="3.573868644s" podCreationTimestamp="2025-10-06 10:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:34.572148365 +0000 UTC m=+5845.362173955" watchObservedRunningTime="2025-10-06 10:16:34.573868644 +0000 UTC m=+5845.363894234" Oct 06 10:16:34 crc kubenswrapper[4989]: I1006 10:16:34.602280 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.602251491 podStartE2EDuration="3.602251491s" podCreationTimestamp="2025-10-06 10:16:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:34.547824035 +0000 UTC m=+5845.337849645" watchObservedRunningTime="2025-10-06 10:16:34.602251491 +0000 UTC m=+5845.392277081" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.104833 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.229915 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.230143 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="dnsmasq-dns" containerID="cri-o://d03d5347d142dfcac09d25cd861f12115a13e1578f1c3c9b7df9007ff48ed6b7" gracePeriod=10 Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.620830 4989 generic.go:334] "Generic (PLEG): container finished" podID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerID="d03d5347d142dfcac09d25cd861f12115a13e1578f1c3c9b7df9007ff48ed6b7" exitCode=0 Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.621150 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" event={"ID":"1ff41850-0386-4b7e-98b1-2fff72b7b784","Type":"ContainerDied","Data":"d03d5347d142dfcac09d25cd861f12115a13e1578f1c3c9b7df9007ff48ed6b7"} Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.703807 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.728902 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config\") pod \"1ff41850-0386-4b7e-98b1-2fff72b7b784\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.729012 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggz7m\" (UniqueName: \"kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m\") pod \"1ff41850-0386-4b7e-98b1-2fff72b7b784\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.729069 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc\") pod \"1ff41850-0386-4b7e-98b1-2fff72b7b784\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.729108 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb\") pod \"1ff41850-0386-4b7e-98b1-2fff72b7b784\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.729192 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb\") pod \"1ff41850-0386-4b7e-98b1-2fff72b7b784\" (UID: \"1ff41850-0386-4b7e-98b1-2fff72b7b784\") " Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.756851 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m" (OuterVolumeSpecName: "kube-api-access-ggz7m") pod "1ff41850-0386-4b7e-98b1-2fff72b7b784" (UID: "1ff41850-0386-4b7e-98b1-2fff72b7b784"). InnerVolumeSpecName "kube-api-access-ggz7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.798205 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config" (OuterVolumeSpecName: "config") pod "1ff41850-0386-4b7e-98b1-2fff72b7b784" (UID: "1ff41850-0386-4b7e-98b1-2fff72b7b784"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.836186 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1ff41850-0386-4b7e-98b1-2fff72b7b784" (UID: "1ff41850-0386-4b7e-98b1-2fff72b7b784"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.837757 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ff41850-0386-4b7e-98b1-2fff72b7b784" (UID: "1ff41850-0386-4b7e-98b1-2fff72b7b784"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.838125 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.838145 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggz7m\" (UniqueName: \"kubernetes.io/projected/1ff41850-0386-4b7e-98b1-2fff72b7b784-kube-api-access-ggz7m\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.838155 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.838165 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.856783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1ff41850-0386-4b7e-98b1-2fff72b7b784" (UID: "1ff41850-0386-4b7e-98b1-2fff72b7b784"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:16:42 crc kubenswrapper[4989]: I1006 10:16:42.939375 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ff41850-0386-4b7e-98b1-2fff72b7b784-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.636883 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" event={"ID":"1ff41850-0386-4b7e-98b1-2fff72b7b784","Type":"ContainerDied","Data":"b19302938f5d02eb47a8424e34275436a59f81eec715cef255662a028f778641"} Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.636948 4989 scope.go:117] "RemoveContainer" containerID="d03d5347d142dfcac09d25cd861f12115a13e1578f1c3c9b7df9007ff48ed6b7" Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.636997 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845cdfbdfc-8jssc" Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.674805 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.677268 4989 scope.go:117] "RemoveContainer" containerID="a6c978abbacda0a19a12c002a7cc830f2778e9490bebc33a80f2e691fe8e4656" Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.686959 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845cdfbdfc-8jssc"] Oct 06 10:16:43 crc kubenswrapper[4989]: I1006 10:16:43.947018 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" path="/var/lib/kubelet/pods/1ff41850-0386-4b7e-98b1-2fff72b7b784/volumes" Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.113898 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.114410 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://66dbc734a43dd45eccc68b9de1126961a0923b2713543855cbd7bc540572f70e" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.135127 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.135372 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="50c10569-bfee-46ef-bdf0-bcd8aea02659" containerName="nova-scheduler-scheduler" containerID="cri-o://409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.145465 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.145775 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-log" containerID="cri-o://7e2bba5ca8017406b6db629faa03406fef22830bae3ed86b9eb168efad28cf5a" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.145943 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-api" containerID="cri-o://4d03966f08e652c143bac94ca7afe54a990d2ebe508f8d91c76de6cd995bd4d7" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.161131 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.161455 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" containerID="cri-o://daf18c273de7a1b7a87ec84169ae813b9b36f24ad44f46542fc90923855737db" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.162010 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" containerID="cri-o://11f723384ee2b989d4b8b4e11647be6158de5fa6075da0134eda587a59a800c3" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.170173 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.170430 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="526b003d-1c65-453c-9c1f-31003fa6d0a3" containerName="nova-cell0-conductor-conductor" containerID="cri-o://e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4" gracePeriod=30 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.363877 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.647936 4989 generic.go:334] "Generic (PLEG): container finished" podID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerID="66dbc734a43dd45eccc68b9de1126961a0923b2713543855cbd7bc540572f70e" exitCode=0 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.647990 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1b4c62a-3cae-4649-b041-91fc4cd606f0","Type":"ContainerDied","Data":"66dbc734a43dd45eccc68b9de1126961a0923b2713543855cbd7bc540572f70e"} Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.658678 4989 generic.go:334] "Generic (PLEG): container finished" podID="14a26d46-0479-4e10-92c9-2479a97b5218" containerID="daf18c273de7a1b7a87ec84169ae813b9b36f24ad44f46542fc90923855737db" exitCode=143 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.658852 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerDied","Data":"daf18c273de7a1b7a87ec84169ae813b9b36f24ad44f46542fc90923855737db"} Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.662332 4989 generic.go:334] "Generic (PLEG): container finished" podID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerID="7e2bba5ca8017406b6db629faa03406fef22830bae3ed86b9eb168efad28cf5a" exitCode=143 Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.662384 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerDied","Data":"7e2bba5ca8017406b6db629faa03406fef22830bae3ed86b9eb168efad28cf5a"} Oct 06 10:16:44 crc kubenswrapper[4989]: I1006 10:16:44.709981 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.60:6080/vnc_lite.html\": dial tcp 10.217.1.60:6080: connect: connection refused" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.027765 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.084050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n48zr\" (UniqueName: \"kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr\") pod \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.084277 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data\") pod \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.084385 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle\") pod \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\" (UID: \"a1b4c62a-3cae-4649-b041-91fc4cd606f0\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.090014 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr" (OuterVolumeSpecName: "kube-api-access-n48zr") pod "a1b4c62a-3cae-4649-b041-91fc4cd606f0" (UID: "a1b4c62a-3cae-4649-b041-91fc4cd606f0"). InnerVolumeSpecName "kube-api-access-n48zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.137169 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data" (OuterVolumeSpecName: "config-data") pod "a1b4c62a-3cae-4649-b041-91fc4cd606f0" (UID: "a1b4c62a-3cae-4649-b041-91fc4cd606f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.142859 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1b4c62a-3cae-4649-b041-91fc4cd606f0" (UID: "a1b4c62a-3cae-4649-b041-91fc4cd606f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.155346 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.185392 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle\") pod \"50c10569-bfee-46ef-bdf0-bcd8aea02659\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.185616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s6rw\" (UniqueName: \"kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw\") pod \"50c10569-bfee-46ef-bdf0-bcd8aea02659\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.185715 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data\") pod \"50c10569-bfee-46ef-bdf0-bcd8aea02659\" (UID: \"50c10569-bfee-46ef-bdf0-bcd8aea02659\") " Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.187187 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.187234 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b4c62a-3cae-4649-b041-91fc4cd606f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.187246 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n48zr\" (UniqueName: \"kubernetes.io/projected/a1b4c62a-3cae-4649-b041-91fc4cd606f0-kube-api-access-n48zr\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.192478 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw" (OuterVolumeSpecName: "kube-api-access-5s6rw") pod "50c10569-bfee-46ef-bdf0-bcd8aea02659" (UID: "50c10569-bfee-46ef-bdf0-bcd8aea02659"). InnerVolumeSpecName "kube-api-access-5s6rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.215417 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50c10569-bfee-46ef-bdf0-bcd8aea02659" (UID: "50c10569-bfee-46ef-bdf0-bcd8aea02659"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.228774 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data" (OuterVolumeSpecName: "config-data") pod "50c10569-bfee-46ef-bdf0-bcd8aea02659" (UID: "50c10569-bfee-46ef-bdf0-bcd8aea02659"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.288802 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s6rw\" (UniqueName: \"kubernetes.io/projected/50c10569-bfee-46ef-bdf0-bcd8aea02659-kube-api-access-5s6rw\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.288837 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.288847 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c10569-bfee-46ef-bdf0-bcd8aea02659-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.674998 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a1b4c62a-3cae-4649-b041-91fc4cd606f0","Type":"ContainerDied","Data":"af4936eaeca81ee6358c02896e480191c567bb48f8b9598074b4f80924a58966"} Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.675048 4989 scope.go:117] "RemoveContainer" containerID="66dbc734a43dd45eccc68b9de1126961a0923b2713543855cbd7bc540572f70e" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.675142 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.682409 4989 generic.go:334] "Generic (PLEG): container finished" podID="50c10569-bfee-46ef-bdf0-bcd8aea02659" containerID="409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2" exitCode=0 Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.682451 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50c10569-bfee-46ef-bdf0-bcd8aea02659","Type":"ContainerDied","Data":"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2"} Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.682480 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"50c10569-bfee-46ef-bdf0-bcd8aea02659","Type":"ContainerDied","Data":"7e3801f61d4fc90fb1938f06abd10f882a1713b5fdb9644ffb070d8ea098bae4"} Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.682487 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.727679 4989 scope.go:117] "RemoveContainer" containerID="409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.759855 4989 scope.go:117] "RemoveContainer" containerID="409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2" Oct 06 10:16:45 crc kubenswrapper[4989]: E1006 10:16:45.760404 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2\": container with ID starting with 409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2 not found: ID does not exist" containerID="409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.760438 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2"} err="failed to get container status \"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2\": rpc error: code = NotFound desc = could not find container \"409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2\": container with ID starting with 409bfe8c3ccceb8665adc4d563d15f2ab7a1a09c753e1b5fe9b0da1bec26d8a2 not found: ID does not exist" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.764150 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.776057 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.794433 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.807118 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.815636 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: E1006 10:16:45.816058 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c10569-bfee-46ef-bdf0-bcd8aea02659" containerName="nova-scheduler-scheduler" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816077 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c10569-bfee-46ef-bdf0-bcd8aea02659" containerName="nova-scheduler-scheduler" Oct 06 10:16:45 crc kubenswrapper[4989]: E1006 10:16:45.816103 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="init" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816109 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="init" Oct 06 10:16:45 crc kubenswrapper[4989]: E1006 10:16:45.816129 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816135 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:16:45 crc kubenswrapper[4989]: E1006 10:16:45.816142 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816148 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816316 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c10569-bfee-46ef-bdf0-bcd8aea02659" containerName="nova-scheduler-scheduler" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816333 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" containerName="nova-cell1-novncproxy-novncproxy" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816349 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ff41850-0386-4b7e-98b1-2fff72b7b784" containerName="dnsmasq-dns" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.816957 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.821511 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.826271 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.828427 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.834460 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.839208 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.851323 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.899053 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfmb\" (UniqueName: \"kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.899201 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.899235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.899299 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.899369 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:45 crc kubenswrapper[4989]: I1006 10:16:45.906435 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5jg\" (UniqueName: \"kubernetes.io/projected/82f58534-c92c-4367-a1ca-1dfcef75384b-kube-api-access-7r5jg\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.017796 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c10569-bfee-46ef-bdf0-bcd8aea02659" path="/var/lib/kubelet/pods/50c10569-bfee-46ef-bdf0-bcd8aea02659/volumes" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.018883 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b4c62a-3cae-4649-b041-91fc4cd606f0" path="/var/lib/kubelet/pods/a1b4c62a-3cae-4649-b041-91fc4cd606f0/volumes" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.029983 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpfmb\" (UniqueName: \"kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.030086 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.030120 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.030170 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.030206 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.030269 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5jg\" (UniqueName: \"kubernetes.io/projected/82f58534-c92c-4367-a1ca-1dfcef75384b-kube-api-access-7r5jg\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.042307 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.047362 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.057187 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.057597 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f58534-c92c-4367-a1ca-1dfcef75384b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.057996 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpfmb\" (UniqueName: \"kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb\") pod \"nova-scheduler-0\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.074385 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5jg\" (UniqueName: \"kubernetes.io/projected/82f58534-c92c-4367-a1ca-1dfcef75384b-kube-api-access-7r5jg\") pod \"nova-cell1-novncproxy-0\" (UID: \"82f58534-c92c-4367-a1ca-1dfcef75384b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.134907 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.149868 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.612979 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 10:16:46 crc kubenswrapper[4989]: W1006 10:16:46.620042 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f8ba4f7_dc15_4ba2_af9a_8f9faba12a15.slice/crio-dcb2dd865fe7fe2bd76e3993478ac09c3728da938247ead7cc59e8480e0d2f24 WatchSource:0}: Error finding container dcb2dd865fe7fe2bd76e3993478ac09c3728da938247ead7cc59e8480e0d2f24: Status 404 returned error can't find the container with id dcb2dd865fe7fe2bd76e3993478ac09c3728da938247ead7cc59e8480e0d2f24 Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.662057 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 06 10:16:46 crc kubenswrapper[4989]: I1006 10:16:46.702172 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15","Type":"ContainerStarted","Data":"dcb2dd865fe7fe2bd76e3993478ac09c3728da938247ead7cc59e8480e0d2f24"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.338023 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": read tcp 10.217.0.2:43070->10.217.1.70:8775: read: connection reset by peer" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.338035 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.70:8775/\": read tcp 10.217.0.2:43080->10.217.1.70:8775: read: connection reset by peer" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.428385 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.428601 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="557be752-c1ce-49a2-8cc0-7f73305c66b5" containerName="nova-cell1-conductor-conductor" containerID="cri-o://9d75d55d504f174a69d3c97fb792d6cad40c23a296dc43d69448a5835f8d3c23" gracePeriod=30 Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.714493 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82f58534-c92c-4367-a1ca-1dfcef75384b","Type":"ContainerStarted","Data":"f6f23d5ecda543c82f6a76c1e3b98949a1a6e8315bcf5352dbfbe4933d415b45"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.715128 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"82f58534-c92c-4367-a1ca-1dfcef75384b","Type":"ContainerStarted","Data":"fa7b869110f30bb94b809ceddce47ddb00817f5396acd5e304f7d61df2da37ac"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.725207 4989 generic.go:334] "Generic (PLEG): container finished" podID="14a26d46-0479-4e10-92c9-2479a97b5218" containerID="11f723384ee2b989d4b8b4e11647be6158de5fa6075da0134eda587a59a800c3" exitCode=0 Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.725266 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerDied","Data":"11f723384ee2b989d4b8b4e11647be6158de5fa6075da0134eda587a59a800c3"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.727500 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.734107 4989 generic.go:334] "Generic (PLEG): container finished" podID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerID="4d03966f08e652c143bac94ca7afe54a990d2ebe508f8d91c76de6cd995bd4d7" exitCode=0 Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.734283 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerDied","Data":"4d03966f08e652c143bac94ca7afe54a990d2ebe508f8d91c76de6cd995bd4d7"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.741336 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15","Type":"ContainerStarted","Data":"34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.744608 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.74457908 podStartE2EDuration="2.74457908s" podCreationTimestamp="2025-10-06 10:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:47.743571721 +0000 UTC m=+5858.533597301" watchObservedRunningTime="2025-10-06 10:16:47.74457908 +0000 UTC m=+5858.534604660" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.760094 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.760300 4989 generic.go:334] "Generic (PLEG): container finished" podID="526b003d-1c65-453c-9c1f-31003fa6d0a3" containerID="e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4" exitCode=0 Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.760381 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"526b003d-1c65-453c-9c1f-31003fa6d0a3","Type":"ContainerDied","Data":"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.760427 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"526b003d-1c65-453c-9c1f-31003fa6d0a3","Type":"ContainerDied","Data":"cf8c0d7e2c619b04960c55b0b459ca223b194c83fb1033267f3b2ce29145a771"} Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.760620 4989 scope.go:117] "RemoveContainer" containerID="e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.768292 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.768264491 podStartE2EDuration="2.768264491s" podCreationTimestamp="2025-10-06 10:16:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:47.766609234 +0000 UTC m=+5858.556634814" watchObservedRunningTime="2025-10-06 10:16:47.768264491 +0000 UTC m=+5858.558290071" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.787376 4989 scope.go:117] "RemoveContainer" containerID="e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.787425 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96qpc\" (UniqueName: \"kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc\") pod \"526b003d-1c65-453c-9c1f-31003fa6d0a3\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " Oct 06 10:16:47 crc kubenswrapper[4989]: E1006 10:16:47.788018 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4\": container with ID starting with e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4 not found: ID does not exist" containerID="e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.788067 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4"} err="failed to get container status \"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4\": rpc error: code = NotFound desc = could not find container \"e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4\": container with ID starting with e6c62fc0f1800031fd84807c64545419c8ae22b1c30cc09786537d6f6bb56eb4 not found: ID does not exist" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.791201 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data\") pod \"526b003d-1c65-453c-9c1f-31003fa6d0a3\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.791256 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle\") pod \"526b003d-1c65-453c-9c1f-31003fa6d0a3\" (UID: \"526b003d-1c65-453c-9c1f-31003fa6d0a3\") " Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.810352 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc" (OuterVolumeSpecName: "kube-api-access-96qpc") pod "526b003d-1c65-453c-9c1f-31003fa6d0a3" (UID: "526b003d-1c65-453c-9c1f-31003fa6d0a3"). InnerVolumeSpecName "kube-api-access-96qpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.831825 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "526b003d-1c65-453c-9c1f-31003fa6d0a3" (UID: "526b003d-1c65-453c-9c1f-31003fa6d0a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.886700 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data" (OuterVolumeSpecName: "config-data") pod "526b003d-1c65-453c-9c1f-31003fa6d0a3" (UID: "526b003d-1c65-453c-9c1f-31003fa6d0a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.897820 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.897856 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526b003d-1c65-453c-9c1f-31003fa6d0a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.897867 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96qpc\" (UniqueName: \"kubernetes.io/projected/526b003d-1c65-453c-9c1f-31003fa6d0a3-kube-api-access-96qpc\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.927229 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:16:47 crc kubenswrapper[4989]: I1006 10:16:47.964308 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.001521 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs\") pod \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.001629 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data\") pod \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.001739 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle\") pod \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.001792 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gppzg\" (UniqueName: \"kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg\") pod \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\" (UID: \"8ab7d5dd-7194-48f0-826d-4c0e986a28a6\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.004421 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs" (OuterVolumeSpecName: "logs") pod "8ab7d5dd-7194-48f0-826d-4c0e986a28a6" (UID: "8ab7d5dd-7194-48f0-826d-4c0e986a28a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.012979 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg" (OuterVolumeSpecName: "kube-api-access-gppzg") pod "8ab7d5dd-7194-48f0-826d-4c0e986a28a6" (UID: "8ab7d5dd-7194-48f0-826d-4c0e986a28a6"). InnerVolumeSpecName "kube-api-access-gppzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.045275 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data" (OuterVolumeSpecName: "config-data") pod "8ab7d5dd-7194-48f0-826d-4c0e986a28a6" (UID: "8ab7d5dd-7194-48f0-826d-4c0e986a28a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.054021 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ab7d5dd-7194-48f0-826d-4c0e986a28a6" (UID: "8ab7d5dd-7194-48f0-826d-4c0e986a28a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.086728 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.097582 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.103480 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data\") pod \"14a26d46-0479-4e10-92c9-2479a97b5218\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.103614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle\") pod \"14a26d46-0479-4e10-92c9-2479a97b5218\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.103766 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xmcz\" (UniqueName: \"kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz\") pod \"14a26d46-0479-4e10-92c9-2479a97b5218\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.103854 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs\") pod \"14a26d46-0479-4e10-92c9-2479a97b5218\" (UID: \"14a26d46-0479-4e10-92c9-2479a97b5218\") " Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.104258 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.104274 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.104285 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gppzg\" (UniqueName: \"kubernetes.io/projected/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-kube-api-access-gppzg\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.104293 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ab7d5dd-7194-48f0-826d-4c0e986a28a6-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.104662 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs" (OuterVolumeSpecName: "logs") pod "14a26d46-0479-4e10-92c9-2479a97b5218" (UID: "14a26d46-0479-4e10-92c9-2479a97b5218"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.106800 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: E1006 10:16:48.107382 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107399 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" Oct 06 10:16:48 crc kubenswrapper[4989]: E1006 10:16:48.107417 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-log" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107423 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-log" Oct 06 10:16:48 crc kubenswrapper[4989]: E1006 10:16:48.107448 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526b003d-1c65-453c-9c1f-31003fa6d0a3" containerName="nova-cell0-conductor-conductor" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107455 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="526b003d-1c65-453c-9c1f-31003fa6d0a3" containerName="nova-cell0-conductor-conductor" Oct 06 10:16:48 crc kubenswrapper[4989]: E1006 10:16:48.107463 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-api" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107471 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-api" Oct 06 10:16:48 crc kubenswrapper[4989]: E1006 10:16:48.107482 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107491 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107703 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-metadata" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107719 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" containerName="nova-metadata-log" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107740 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-log" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107750 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" containerName="nova-api-api" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.107761 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="526b003d-1c65-453c-9c1f-31003fa6d0a3" containerName="nova-cell0-conductor-conductor" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.108637 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.109384 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz" (OuterVolumeSpecName: "kube-api-access-6xmcz") pod "14a26d46-0479-4e10-92c9-2479a97b5218" (UID: "14a26d46-0479-4e10-92c9-2479a97b5218"). InnerVolumeSpecName "kube-api-access-6xmcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.113233 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.135211 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.145533 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14a26d46-0479-4e10-92c9-2479a97b5218" (UID: "14a26d46-0479-4e10-92c9-2479a97b5218"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.154865 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data" (OuterVolumeSpecName: "config-data") pod "14a26d46-0479-4e10-92c9-2479a97b5218" (UID: "14a26d46-0479-4e10-92c9-2479a97b5218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209613 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zddlj\" (UniqueName: \"kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209675 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209765 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209853 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xmcz\" (UniqueName: \"kubernetes.io/projected/14a26d46-0479-4e10-92c9-2479a97b5218-kube-api-access-6xmcz\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209866 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14a26d46-0479-4e10-92c9-2479a97b5218-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209879 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.209889 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a26d46-0479-4e10-92c9-2479a97b5218-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.311505 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.311681 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zddlj\" (UniqueName: \"kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.311715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.316330 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.316459 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.330221 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zddlj\" (UniqueName: \"kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj\") pod \"nova-cell0-conductor-0\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.430961 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.769446 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"14a26d46-0479-4e10-92c9-2479a97b5218","Type":"ContainerDied","Data":"d1a166308e0bc354ea46b758d16497d1187eb20de1b3ec2949cc95043b8f9023"} Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.769802 4989 scope.go:117] "RemoveContainer" containerID="11f723384ee2b989d4b8b4e11647be6158de5fa6075da0134eda587a59a800c3" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.769911 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.775453 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.775905 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8ab7d5dd-7194-48f0-826d-4c0e986a28a6","Type":"ContainerDied","Data":"404626d57d95245f3818a1d98ac9864022ed0f5738e489b1122af399ae7b4ab9"} Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.793217 4989 scope.go:117] "RemoveContainer" containerID="daf18c273de7a1b7a87ec84169ae813b9b36f24ad44f46542fc90923855737db" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.815393 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.849792 4989 scope.go:117] "RemoveContainer" containerID="4d03966f08e652c143bac94ca7afe54a990d2ebe508f8d91c76de6cd995bd4d7" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.889623 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.911141 4989 scope.go:117] "RemoveContainer" containerID="7e2bba5ca8017406b6db629faa03406fef22830bae3ed86b9eb168efad28cf5a" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.912995 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.922155 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.929393 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.930872 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.932735 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.936268 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.945143 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.946927 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.950160 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.956181 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:48 crc kubenswrapper[4989]: I1006 10:16:48.973237 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026097 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026138 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026165 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dhxc\" (UniqueName: \"kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026185 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026218 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026234 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026251 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.026271 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbsz\" (UniqueName: \"kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128299 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128669 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128694 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128721 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbsz\" (UniqueName: \"kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128831 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128847 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128868 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dhxc\" (UniqueName: \"kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.128890 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.129107 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.130079 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.134127 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.134574 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.137850 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.144418 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.147587 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbsz\" (UniqueName: \"kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz\") pod \"nova-api-0\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.149228 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dhxc\" (UniqueName: \"kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc\") pod \"nova-metadata-0\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.383460 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.405119 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.794301 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d03995f-8dfe-424a-a482-213b5d3b5585","Type":"ContainerStarted","Data":"fef29ab3688c3a740816402adee008be8f7c5dd96401d75cd7449b94c3bee53e"} Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.794567 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d03995f-8dfe-424a-a482-213b5d3b5585","Type":"ContainerStarted","Data":"58257d0009f1d426760ceaa52c5d7483782b81bca79776432b7b9d547702f028"} Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.795847 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.819341 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.81932561 podStartE2EDuration="1.81932561s" podCreationTimestamp="2025-10-06 10:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:49.81342521 +0000 UTC m=+5860.603450790" watchObservedRunningTime="2025-10-06 10:16:49.81932561 +0000 UTC m=+5860.609351190" Oct 06 10:16:49 crc kubenswrapper[4989]: W1006 10:16:49.940535 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c24cb4b_6793_4479_86e5_fd42c3a91717.slice/crio-f97986c8414427053b9e48c4b982210157585ec9b85d28f218d6ae22d6e9c026 WatchSource:0}: Error finding container f97986c8414427053b9e48c4b982210157585ec9b85d28f218d6ae22d6e9c026: Status 404 returned error can't find the container with id f97986c8414427053b9e48c4b982210157585ec9b85d28f218d6ae22d6e9c026 Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.959845 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a26d46-0479-4e10-92c9-2479a97b5218" path="/var/lib/kubelet/pods/14a26d46-0479-4e10-92c9-2479a97b5218/volumes" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.960406 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="526b003d-1c65-453c-9c1f-31003fa6d0a3" path="/var/lib/kubelet/pods/526b003d-1c65-453c-9c1f-31003fa6d0a3/volumes" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.963361 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab7d5dd-7194-48f0-826d-4c0e986a28a6" path="/var/lib/kubelet/pods/8ab7d5dd-7194-48f0-826d-4c0e986a28a6/volumes" Oct 06 10:16:49 crc kubenswrapper[4989]: I1006 10:16:49.964054 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.015768 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 10:16:50 crc kubenswrapper[4989]: W1006 10:16:50.026610 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81438bf6_a32a_443e_a94a_b6882c9b8ca3.slice/crio-7559f357d36fb81e79f90d49151a4737fe312b60003d91fcff3604f39262742c WatchSource:0}: Error finding container 7559f357d36fb81e79f90d49151a4737fe312b60003d91fcff3604f39262742c: Status 404 returned error can't find the container with id 7559f357d36fb81e79f90d49151a4737fe312b60003d91fcff3604f39262742c Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.805410 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerStarted","Data":"5545728d23140fb4a357835b2146bb2fd1b83a84a4ec6d677315b34e3932d53c"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.805872 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerStarted","Data":"49715fe8fe724fc16a887d8425cb2688792ec1ef1b047be7fbff4eb82409acd4"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.805895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerStarted","Data":"7559f357d36fb81e79f90d49151a4737fe312b60003d91fcff3604f39262742c"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.808402 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerStarted","Data":"8a8a4a26f64508dd625d868e89d5a5a400480e11183c70bd21c8cc5895e17d2c"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.808439 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerStarted","Data":"3f2141d4fcbcd8cea4a2c0c4c90b8dbe7553a806e33c6d3f7b6c3a9388fdbcfa"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.808449 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerStarted","Data":"f97986c8414427053b9e48c4b982210157585ec9b85d28f218d6ae22d6e9c026"} Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.826865 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.82684719 podStartE2EDuration="2.82684719s" podCreationTimestamp="2025-10-06 10:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:50.821578258 +0000 UTC m=+5861.611603838" watchObservedRunningTime="2025-10-06 10:16:50.82684719 +0000 UTC m=+5861.616872770" Oct 06 10:16:50 crc kubenswrapper[4989]: I1006 10:16:50.845280 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.84525887 podStartE2EDuration="2.84525887s" podCreationTimestamp="2025-10-06 10:16:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:50.840690868 +0000 UTC m=+5861.630716458" watchObservedRunningTime="2025-10-06 10:16:50.84525887 +0000 UTC m=+5861.635284460" Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.135287 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.151080 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.825483 4989 generic.go:334] "Generic (PLEG): container finished" podID="557be752-c1ce-49a2-8cc0-7f73305c66b5" containerID="9d75d55d504f174a69d3c97fb792d6cad40c23a296dc43d69448a5835f8d3c23" exitCode=0 Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.825569 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"557be752-c1ce-49a2-8cc0-7f73305c66b5","Type":"ContainerDied","Data":"9d75d55d504f174a69d3c97fb792d6cad40c23a296dc43d69448a5835f8d3c23"} Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.830074 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"557be752-c1ce-49a2-8cc0-7f73305c66b5","Type":"ContainerDied","Data":"f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784"} Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.830088 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f88f5fb6d23c6fef2e15f4dfbf6e7ed0273da50750b1ad770318841580941784" Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.852791 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.992453 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data\") pod \"557be752-c1ce-49a2-8cc0-7f73305c66b5\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.992522 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtsng\" (UniqueName: \"kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng\") pod \"557be752-c1ce-49a2-8cc0-7f73305c66b5\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " Oct 06 10:16:51 crc kubenswrapper[4989]: I1006 10:16:51.992587 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle\") pod \"557be752-c1ce-49a2-8cc0-7f73305c66b5\" (UID: \"557be752-c1ce-49a2-8cc0-7f73305c66b5\") " Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.013171 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng" (OuterVolumeSpecName: "kube-api-access-mtsng") pod "557be752-c1ce-49a2-8cc0-7f73305c66b5" (UID: "557be752-c1ce-49a2-8cc0-7f73305c66b5"). InnerVolumeSpecName "kube-api-access-mtsng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.027767 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "557be752-c1ce-49a2-8cc0-7f73305c66b5" (UID: "557be752-c1ce-49a2-8cc0-7f73305c66b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.050062 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data" (OuterVolumeSpecName: "config-data") pod "557be752-c1ce-49a2-8cc0-7f73305c66b5" (UID: "557be752-c1ce-49a2-8cc0-7f73305c66b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.094829 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.094874 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtsng\" (UniqueName: \"kubernetes.io/projected/557be752-c1ce-49a2-8cc0-7f73305c66b5-kube-api-access-mtsng\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.094886 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/557be752-c1ce-49a2-8cc0-7f73305c66b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.839200 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.909981 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.917830 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.948642 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:52 crc kubenswrapper[4989]: E1006 10:16:52.949058 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557be752-c1ce-49a2-8cc0-7f73305c66b5" containerName="nova-cell1-conductor-conductor" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.949077 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="557be752-c1ce-49a2-8cc0-7f73305c66b5" containerName="nova-cell1-conductor-conductor" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.949619 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="557be752-c1ce-49a2-8cc0-7f73305c66b5" containerName="nova-cell1-conductor-conductor" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.950709 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.953205 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 10:16:52 crc kubenswrapper[4989]: I1006 10:16:52.959832 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.113006 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnf7b\" (UniqueName: \"kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.113075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.113154 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.215167 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnf7b\" (UniqueName: \"kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.215266 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.215389 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.223887 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.224072 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.247235 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnf7b\" (UniqueName: \"kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b\") pod \"nova-cell1-conductor-0\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.271427 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.596272 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 10:16:53 crc kubenswrapper[4989]: W1006 10:16:53.603359 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93484eea_55ba_414b_b7e2_a969439587f4.slice/crio-04af45d9118e580b61e613a1eeb7bba118ea0db6a3245cf05bc00fe8f02a51d2 WatchSource:0}: Error finding container 04af45d9118e580b61e613a1eeb7bba118ea0db6a3245cf05bc00fe8f02a51d2: Status 404 returned error can't find the container with id 04af45d9118e580b61e613a1eeb7bba118ea0db6a3245cf05bc00fe8f02a51d2 Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.852749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93484eea-55ba-414b-b7e2-a969439587f4","Type":"ContainerStarted","Data":"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db"} Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.853152 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93484eea-55ba-414b-b7e2-a969439587f4","Type":"ContainerStarted","Data":"04af45d9118e580b61e613a1eeb7bba118ea0db6a3245cf05bc00fe8f02a51d2"} Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.853174 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.894364 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.8943463139999999 podStartE2EDuration="1.894346314s" podCreationTimestamp="2025-10-06 10:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:16:53.888205789 +0000 UTC m=+5864.678231379" watchObservedRunningTime="2025-10-06 10:16:53.894346314 +0000 UTC m=+5864.684371884" Oct 06 10:16:53 crc kubenswrapper[4989]: I1006 10:16:53.948247 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="557be752-c1ce-49a2-8cc0-7f73305c66b5" path="/var/lib/kubelet/pods/557be752-c1ce-49a2-8cc0-7f73305c66b5/volumes" Oct 06 10:16:54 crc kubenswrapper[4989]: I1006 10:16:54.383891 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:16:54 crc kubenswrapper[4989]: I1006 10:16:54.383946 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.135533 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.153396 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.172069 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.185639 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.915942 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 06 10:16:56 crc kubenswrapper[4989]: I1006 10:16:56.942935 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 10:16:58 crc kubenswrapper[4989]: I1006 10:16:58.298966 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 06 10:16:58 crc kubenswrapper[4989]: I1006 10:16:58.468407 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 06 10:16:59 crc kubenswrapper[4989]: I1006 10:16:59.384199 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:16:59 crc kubenswrapper[4989]: I1006 10:16:59.386932 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 10:16:59 crc kubenswrapper[4989]: I1006 10:16:59.405855 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:16:59 crc kubenswrapper[4989]: I1006 10:16:59.405922 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 10:17:00 crc kubenswrapper[4989]: I1006 10:17:00.547902 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:17:00 crc kubenswrapper[4989]: I1006 10:17:00.547910 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:17:00 crc kubenswrapper[4989]: I1006 10:17:00.547940 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:17:00 crc kubenswrapper[4989]: I1006 10:17:00.548192 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.82:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.496791 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.499028 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.502693 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.531378 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626045 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626307 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626400 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf54b\" (UniqueName: \"kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626633 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.626741 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.728788 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.729274 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.729566 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.729381 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.730738 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf54b\" (UniqueName: \"kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.731045 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.731230 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.737719 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.738335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.738842 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.740960 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.752752 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf54b\" (UniqueName: \"kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b\") pod \"cinder-scheduler-0\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:03 crc kubenswrapper[4989]: I1006 10:17:03.822592 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:04 crc kubenswrapper[4989]: I1006 10:17:04.302719 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:04 crc kubenswrapper[4989]: W1006 10:17:04.303740 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0713909_2cc4_4041_b4c9_f9ceace45bc2.slice/crio-b59c2648f4aaa870086a678dfcf4918e7226fc566b3fcfe3913538613657b8cb WatchSource:0}: Error finding container b59c2648f4aaa870086a678dfcf4918e7226fc566b3fcfe3913538613657b8cb: Status 404 returned error can't find the container with id b59c2648f4aaa870086a678dfcf4918e7226fc566b3fcfe3913538613657b8cb Oct 06 10:17:04 crc kubenswrapper[4989]: I1006 10:17:04.986331 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerStarted","Data":"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb"} Oct 06 10:17:04 crc kubenswrapper[4989]: I1006 10:17:04.986890 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerStarted","Data":"b59c2648f4aaa870086a678dfcf4918e7226fc566b3fcfe3913538613657b8cb"} Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.162697 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.162998 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api-log" containerID="cri-o://cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235" gracePeriod=30 Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.163476 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api" containerID="cri-o://3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6" gracePeriod=30 Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.853443 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.856540 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.859763 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.911860 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984749 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984836 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984862 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984909 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984950 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-run\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.984985 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985010 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985042 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985073 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlprc\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-kube-api-access-rlprc\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985097 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985120 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985140 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985162 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985191 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985232 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:05 crc kubenswrapper[4989]: I1006 10:17:05.985258 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.013578 4989 generic.go:334] "Generic (PLEG): container finished" podID="264b1db1-3250-43e1-9928-04cd00bb5052" containerID="cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235" exitCode=143 Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.013701 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerDied","Data":"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235"} Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.018551 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerStarted","Data":"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc"} Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.033443 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.033411545 podStartE2EDuration="3.033411545s" podCreationTimestamp="2025-10-06 10:17:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:06.031274594 +0000 UTC m=+5876.821300174" watchObservedRunningTime="2025-10-06 10:17:06.033411545 +0000 UTC m=+5876.823437125" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089808 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089876 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-run\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089908 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089928 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089951 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089972 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlprc\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-kube-api-access-rlprc\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.089990 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090009 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090024 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090037 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090061 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090087 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090107 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090130 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090250 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090269 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.090413 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.097491 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.097569 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-run\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.097790 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.097831 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098110 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098150 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098170 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098374 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098464 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.098502 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0fbd8084-52d9-4126-a357-c6e07168a716-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.105814 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.107621 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.111999 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.112468 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fbd8084-52d9-4126-a357-c6e07168a716-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.117578 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlprc\" (UniqueName: \"kubernetes.io/projected/0fbd8084-52d9-4126-a357-c6e07168a716-kube-api-access-rlprc\") pod \"cinder-volume-volume1-0\" (UID: \"0fbd8084-52d9-4126-a357-c6e07168a716\") " pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.216975 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.364726 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.367476 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.372196 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.375103 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497062 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-sys\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497456 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-ceph\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497503 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-run\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497528 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497546 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497564 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497585 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-lib-modules\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497604 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data-custom\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497635 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqz9v\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-kube-api-access-fqz9v\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497670 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497689 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497703 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-dev\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497731 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497791 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-nvme\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.497808 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-scripts\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599849 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599895 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-lib-modules\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599939 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data-custom\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599959 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqz9v\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-kube-api-access-fqz9v\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.599994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-dev\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600008 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600036 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600105 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-nvme\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600123 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-scripts\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600149 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-sys\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600164 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600278 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-ceph\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600314 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-run\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.600337 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601254 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601519 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601558 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-lib-modules\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601591 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-dev\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601638 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-sys\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601647 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601726 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-nvme\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601741 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.601776 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/67780570-63ad-40e2-88c9-dc4bb2563723-run\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.607470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-scripts\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.607962 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.608421 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.618124 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-ceph\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.618296 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67780570-63ad-40e2-88c9-dc4bb2563723-config-data-custom\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.621271 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqz9v\" (UniqueName: \"kubernetes.io/projected/67780570-63ad-40e2-88c9-dc4bb2563723-kube-api-access-fqz9v\") pod \"cinder-backup-0\" (UID: \"67780570-63ad-40e2-88c9-dc4bb2563723\") " pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.697748 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 06 10:17:06 crc kubenswrapper[4989]: I1006 10:17:06.896681 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 06 10:17:06 crc kubenswrapper[4989]: W1006 10:17:06.902957 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fbd8084_52d9_4126_a357_c6e07168a716.slice/crio-0c1be13dc1d5ab4b0067d601739ab24f8764817277b72e5c2f50412b4cc75a61 WatchSource:0}: Error finding container 0c1be13dc1d5ab4b0067d601739ab24f8764817277b72e5c2f50412b4cc75a61: Status 404 returned error can't find the container with id 0c1be13dc1d5ab4b0067d601739ab24f8764817277b72e5c2f50412b4cc75a61 Oct 06 10:17:07 crc kubenswrapper[4989]: I1006 10:17:07.029280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0fbd8084-52d9-4126-a357-c6e07168a716","Type":"ContainerStarted","Data":"0c1be13dc1d5ab4b0067d601739ab24f8764817277b72e5c2f50412b4cc75a61"} Oct 06 10:17:07 crc kubenswrapper[4989]: I1006 10:17:07.272698 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.044324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0fbd8084-52d9-4126-a357-c6e07168a716","Type":"ContainerStarted","Data":"944c22e76f1cf2ffb9876f8c8de7f6ac98633441d65bf353bf9b0d99b3a75b8a"} Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.044928 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0fbd8084-52d9-4126-a357-c6e07168a716","Type":"ContainerStarted","Data":"1aca6e5c548e4c95c890806abbedc432e7e81ab3494883fa1e69f606d6c25b9f"} Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.045885 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"67780570-63ad-40e2-88c9-dc4bb2563723","Type":"ContainerStarted","Data":"b1032451a26f0b0be58384f0e820572ddad622d664c123fdde30202fed4ac83d"} Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.077946 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.39203782 podStartE2EDuration="3.077928103s" podCreationTimestamp="2025-10-06 10:17:05 +0000 UTC" firstStartedPulling="2025-10-06 10:17:06.905171645 +0000 UTC m=+5877.695197225" lastFinishedPulling="2025-10-06 10:17:07.591061918 +0000 UTC m=+5878.381087508" observedRunningTime="2025-10-06 10:17:08.071749944 +0000 UTC m=+5878.861775534" watchObservedRunningTime="2025-10-06 10:17:08.077928103 +0000 UTC m=+5878.867953683" Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.512089 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.77:8776/healthcheck\": read tcp 10.217.0.2:33554->10.217.1.77:8776: read: connection reset by peer" Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.823752 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 10:17:08 crc kubenswrapper[4989]: I1006 10:17:08.889615 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.056711 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhzwq\" (UniqueName: \"kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057143 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057166 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057182 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057201 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id\") pod \"264b1db1-3250-43e1-9928-04cd00bb5052\" (UID: \"264b1db1-3250-43e1-9928-04cd00bb5052\") " Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.057637 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.058566 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs" (OuterVolumeSpecName: "logs") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.060940 4989 generic.go:334] "Generic (PLEG): container finished" podID="264b1db1-3250-43e1-9928-04cd00bb5052" containerID="3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6" exitCode=0 Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.061010 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerDied","Data":"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6"} Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.061036 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"264b1db1-3250-43e1-9928-04cd00bb5052","Type":"ContainerDied","Data":"2762ec505277cdd98dcd3f14f068e3db71b1024af2589622a9650d54e561fb89"} Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.061053 4989 scope.go:117] "RemoveContainer" containerID="3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.061057 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.063458 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.065408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"67780570-63ad-40e2-88c9-dc4bb2563723","Type":"ContainerStarted","Data":"a01e8c2fe9da0f8ed49cb92d625936f3aa5f51edb7c524555e473bc4c2dd8308"} Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.066185 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq" (OuterVolumeSpecName: "kube-api-access-qhzwq") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "kube-api-access-qhzwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.068858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts" (OuterVolumeSpecName: "scripts") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.088844 4989 scope.go:117] "RemoveContainer" containerID="cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.107280 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.118149 4989 scope.go:117] "RemoveContainer" containerID="3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.118458 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data" (OuterVolumeSpecName: "config-data") pod "264b1db1-3250-43e1-9928-04cd00bb5052" (UID: "264b1db1-3250-43e1-9928-04cd00bb5052"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:09 crc kubenswrapper[4989]: E1006 10:17:09.118512 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6\": container with ID starting with 3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6 not found: ID does not exist" containerID="3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.118545 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6"} err="failed to get container status \"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6\": rpc error: code = NotFound desc = could not find container \"3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6\": container with ID starting with 3501eda900b6a143de27455e716f50df362b5172cfedf5db2c1b65c739a419c6 not found: ID does not exist" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.118569 4989 scope.go:117] "RemoveContainer" containerID="cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235" Oct 06 10:17:09 crc kubenswrapper[4989]: E1006 10:17:09.118844 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235\": container with ID starting with cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235 not found: ID does not exist" containerID="cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.118882 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235"} err="failed to get container status \"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235\": rpc error: code = NotFound desc = could not find container \"cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235\": container with ID starting with cb1c8acdf7157ebdfbecf3712cc4ee66a025bedd3617c6177b8ecb0e3bea6235 not found: ID does not exist" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160784 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160836 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160851 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhzwq\" (UniqueName: \"kubernetes.io/projected/264b1db1-3250-43e1-9928-04cd00bb5052-kube-api-access-qhzwq\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160867 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/264b1db1-3250-43e1-9928-04cd00bb5052-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160879 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160891 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/264b1db1-3250-43e1-9928-04cd00bb5052-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.160900 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/264b1db1-3250-43e1-9928-04cd00bb5052-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.387156 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.390577 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.396299 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.414909 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.415635 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.416613 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.424517 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.441107 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.441216 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.481949 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:09 crc kubenswrapper[4989]: E1006 10:17:09.482667 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.482689 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api" Oct 06 10:17:09 crc kubenswrapper[4989]: E1006 10:17:09.482743 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api-log" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.482750 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api-log" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.483056 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api-log" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.483086 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" containerName="cinder-api" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.487372 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.490014 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.512811 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.687806 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-scripts\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.687879 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.688272 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9flc\" (UniqueName: \"kubernetes.io/projected/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-kube-api-access-h9flc\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.688357 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.688387 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-logs\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.688416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.688457 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data-custom\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791150 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-logs\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791202 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791237 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data-custom\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791313 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-scripts\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791321 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791400 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.791594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9flc\" (UniqueName: \"kubernetes.io/projected/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-kube-api-access-h9flc\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.792881 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-logs\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.797857 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data-custom\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.807363 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.817156 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-scripts\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.817212 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-config-data\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.820540 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9flc\" (UniqueName: \"kubernetes.io/projected/10ad6b44-28dc-4c46-bb6f-f3c6b898e334-kube-api-access-h9flc\") pod \"cinder-api-0\" (UID: \"10ad6b44-28dc-4c46-bb6f-f3c6b898e334\") " pod="openstack/cinder-api-0" Oct 06 10:17:09 crc kubenswrapper[4989]: I1006 10:17:09.954452 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="264b1db1-3250-43e1-9928-04cd00bb5052" path="/var/lib/kubelet/pods/264b1db1-3250-43e1-9928-04cd00bb5052/volumes" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.075931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"67780570-63ad-40e2-88c9-dc4bb2563723","Type":"ContainerStarted","Data":"e7effc652523ff81961a0ee095c7da3442eb7e35aac0af870c1f2f8705aa16b9"} Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.081062 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.083797 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.083867 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.106173 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.108503 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.926990488 podStartE2EDuration="4.108488176s" podCreationTimestamp="2025-10-06 10:17:06 +0000 UTC" firstStartedPulling="2025-10-06 10:17:07.280706861 +0000 UTC m=+5878.070732441" lastFinishedPulling="2025-10-06 10:17:08.462204549 +0000 UTC m=+5879.252230129" observedRunningTime="2025-10-06 10:17:10.1027225 +0000 UTC m=+5880.892748090" watchObservedRunningTime="2025-10-06 10:17:10.108488176 +0000 UTC m=+5880.898513756" Oct 06 10:17:10 crc kubenswrapper[4989]: I1006 10:17:10.659776 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 06 10:17:10 crc kubenswrapper[4989]: W1006 10:17:10.666169 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10ad6b44_28dc_4c46_bb6f_f3c6b898e334.slice/crio-6bf00d619201b71dbbd37410bcb130507601e6b208067fbce80239ccad673a8a WatchSource:0}: Error finding container 6bf00d619201b71dbbd37410bcb130507601e6b208067fbce80239ccad673a8a: Status 404 returned error can't find the container with id 6bf00d619201b71dbbd37410bcb130507601e6b208067fbce80239ccad673a8a Oct 06 10:17:11 crc kubenswrapper[4989]: I1006 10:17:11.093737 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10ad6b44-28dc-4c46-bb6f-f3c6b898e334","Type":"ContainerStarted","Data":"6bf00d619201b71dbbd37410bcb130507601e6b208067fbce80239ccad673a8a"} Oct 06 10:17:11 crc kubenswrapper[4989]: I1006 10:17:11.217031 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:11 crc kubenswrapper[4989]: I1006 10:17:11.698694 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 06 10:17:12 crc kubenswrapper[4989]: I1006 10:17:12.104472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10ad6b44-28dc-4c46-bb6f-f3c6b898e334","Type":"ContainerStarted","Data":"0a22e0fc57da6461f274cb87cc39f6f28e6961b526f636a3f74ac8003d0ee94c"} Oct 06 10:17:12 crc kubenswrapper[4989]: I1006 10:17:12.105390 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 06 10:17:12 crc kubenswrapper[4989]: I1006 10:17:12.105420 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10ad6b44-28dc-4c46-bb6f-f3c6b898e334","Type":"ContainerStarted","Data":"0f289e6318e0a431442a14913e9d92cb79227c2fac438cfffe03222b79163c8e"} Oct 06 10:17:12 crc kubenswrapper[4989]: I1006 10:17:12.123318 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.123289176 podStartE2EDuration="3.123289176s" podCreationTimestamp="2025-10-06 10:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:12.120600769 +0000 UTC m=+5882.910626349" watchObservedRunningTime="2025-10-06 10:17:12.123289176 +0000 UTC m=+5882.913314756" Oct 06 10:17:14 crc kubenswrapper[4989]: I1006 10:17:14.027699 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 10:17:14 crc kubenswrapper[4989]: I1006 10:17:14.102773 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:14 crc kubenswrapper[4989]: I1006 10:17:14.124746 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="cinder-scheduler" containerID="cri-o://89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb" gracePeriod=30 Oct 06 10:17:14 crc kubenswrapper[4989]: I1006 10:17:14.124791 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="probe" containerID="cri-o://551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc" gracePeriod=30 Oct 06 10:17:15 crc kubenswrapper[4989]: I1006 10:17:15.135972 4989 generic.go:334] "Generic (PLEG): container finished" podID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerID="551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc" exitCode=0 Oct 06 10:17:15 crc kubenswrapper[4989]: I1006 10:17:15.136058 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerDied","Data":"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc"} Oct 06 10:17:16 crc kubenswrapper[4989]: I1006 10:17:16.438102 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 06 10:17:16 crc kubenswrapper[4989]: I1006 10:17:16.907194 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 06 10:17:16 crc kubenswrapper[4989]: I1006 10:17:16.919968 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.050262 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf54b\" (UniqueName: \"kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.050708 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.050848 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.051018 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.051158 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.051234 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.051331 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts\") pod \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\" (UID: \"d0713909-2cc4-4041-b4c9-f9ceace45bc2\") " Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.052042 4989 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d0713909-2cc4-4041-b4c9-f9ceace45bc2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.055847 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts" (OuterVolumeSpecName: "scripts") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.056118 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b" (OuterVolumeSpecName: "kube-api-access-mf54b") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "kube-api-access-mf54b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.064920 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.117380 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.147810 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data" (OuterVolumeSpecName: "config-data") pod "d0713909-2cc4-4041-b4c9-f9ceace45bc2" (UID: "d0713909-2cc4-4041-b4c9-f9ceace45bc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.153927 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf54b\" (UniqueName: \"kubernetes.io/projected/d0713909-2cc4-4041-b4c9-f9ceace45bc2-kube-api-access-mf54b\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.153958 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.153968 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.153979 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.153988 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0713909-2cc4-4041-b4c9-f9ceace45bc2-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.157780 4989 generic.go:334] "Generic (PLEG): container finished" podID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerID="89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb" exitCode=0 Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.157816 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerDied","Data":"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb"} Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.157841 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d0713909-2cc4-4041-b4c9-f9ceace45bc2","Type":"ContainerDied","Data":"b59c2648f4aaa870086a678dfcf4918e7226fc566b3fcfe3913538613657b8cb"} Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.157858 4989 scope.go:117] "RemoveContainer" containerID="551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.157963 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.184947 4989 scope.go:117] "RemoveContainer" containerID="89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.194910 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.211933 4989 scope.go:117] "RemoveContainer" containerID="551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc" Oct 06 10:17:17 crc kubenswrapper[4989]: E1006 10:17:17.212498 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc\": container with ID starting with 551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc not found: ID does not exist" containerID="551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.212534 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc"} err="failed to get container status \"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc\": rpc error: code = NotFound desc = could not find container \"551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc\": container with ID starting with 551dc1a03bbf70b505a750c3ef1c30c03db01f517cd83ecbb3906fca381870dc not found: ID does not exist" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.212561 4989 scope.go:117] "RemoveContainer" containerID="89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb" Oct 06 10:17:17 crc kubenswrapper[4989]: E1006 10:17:17.213001 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb\": container with ID starting with 89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb not found: ID does not exist" containerID="89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.213035 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb"} err="failed to get container status \"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb\": rpc error: code = NotFound desc = could not find container \"89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb\": container with ID starting with 89d91afff657d1d9144f3afd471a60276155fcace30ae8527ff9781d4525cdeb not found: ID does not exist" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.218643 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.227623 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:17 crc kubenswrapper[4989]: E1006 10:17:17.228137 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="probe" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.228161 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="probe" Oct 06 10:17:17 crc kubenswrapper[4989]: E1006 10:17:17.228189 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="cinder-scheduler" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.228196 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="cinder-scheduler" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.228406 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="probe" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.228431 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" containerName="cinder-scheduler" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.229564 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.231222 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.234846 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.356838 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.357220 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.357286 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.357397 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5zbz\" (UniqueName: \"kubernetes.io/projected/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-kube-api-access-c5zbz\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.357593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.357706 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.459689 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.459768 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.459840 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.460050 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.460086 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.460136 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5zbz\" (UniqueName: \"kubernetes.io/projected/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-kube-api-access-c5zbz\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.461367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.466370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.467649 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-scripts\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.469108 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.482199 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-config-data\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.486605 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5zbz\" (UniqueName: \"kubernetes.io/projected/9e92a003-5bf6-4ee8-993f-cc75ee8bd077-kube-api-access-c5zbz\") pod \"cinder-scheduler-0\" (UID: \"9e92a003-5bf6-4ee8-993f-cc75ee8bd077\") " pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.552627 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.854198 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 06 10:17:17 crc kubenswrapper[4989]: W1006 10:17:17.864216 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e92a003_5bf6_4ee8_993f_cc75ee8bd077.slice/crio-c01070419d8350f9a454a23b6fff5b9ccf472f198ab0ab37b5ccd8e9b15dce08 WatchSource:0}: Error finding container c01070419d8350f9a454a23b6fff5b9ccf472f198ab0ab37b5ccd8e9b15dce08: Status 404 returned error can't find the container with id c01070419d8350f9a454a23b6fff5b9ccf472f198ab0ab37b5ccd8e9b15dce08 Oct 06 10:17:17 crc kubenswrapper[4989]: I1006 10:17:17.958415 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0713909-2cc4-4041-b4c9-f9ceace45bc2" path="/var/lib/kubelet/pods/d0713909-2cc4-4041-b4c9-f9ceace45bc2/volumes" Oct 06 10:17:18 crc kubenswrapper[4989]: I1006 10:17:18.169562 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e92a003-5bf6-4ee8-993f-cc75ee8bd077","Type":"ContainerStarted","Data":"c01070419d8350f9a454a23b6fff5b9ccf472f198ab0ab37b5ccd8e9b15dce08"} Oct 06 10:17:19 crc kubenswrapper[4989]: I1006 10:17:19.198923 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e92a003-5bf6-4ee8-993f-cc75ee8bd077","Type":"ContainerStarted","Data":"2fbf6609c8a4fbc474d530af8e6cab38f7620b9785cc34dd5f5d95c33980e5c2"} Oct 06 10:17:19 crc kubenswrapper[4989]: I1006 10:17:19.199594 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9e92a003-5bf6-4ee8-993f-cc75ee8bd077","Type":"ContainerStarted","Data":"45ab087399f2a3a2ac6afd441441e60af24af19223a3f01dd3ff8b077912dc34"} Oct 06 10:17:21 crc kubenswrapper[4989]: I1006 10:17:21.204035 4989 scope.go:117] "RemoveContainer" containerID="bdbbd9375919c80d764b03c9b8cc82a9429e5914a0294927909d2cf84261e0bd" Oct 06 10:17:21 crc kubenswrapper[4989]: I1006 10:17:21.250232 4989 scope.go:117] "RemoveContainer" containerID="5e7d035e7a8d78246173cd60a429bf923ae69673a36dbad70fc327a5bb20ffaa" Oct 06 10:17:21 crc kubenswrapper[4989]: I1006 10:17:21.782193 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 06 10:17:21 crc kubenswrapper[4989]: I1006 10:17:21.812802 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.812768913 podStartE2EDuration="4.812768913s" podCreationTimestamp="2025-10-06 10:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:17:19.225195171 +0000 UTC m=+5890.015220751" watchObservedRunningTime="2025-10-06 10:17:21.812768913 +0000 UTC m=+5892.602794563" Oct 06 10:17:22 crc kubenswrapper[4989]: I1006 10:17:22.552805 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 06 10:17:27 crc kubenswrapper[4989]: I1006 10:17:27.785529 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.329443 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.334211 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.347046 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.432117 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.432252 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.432363 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwpzb\" (UniqueName: \"kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.534105 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.534300 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.534397 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwpzb\" (UniqueName: \"kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.534646 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.534726 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.569662 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwpzb\" (UniqueName: \"kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb\") pod \"redhat-marketplace-7fqcn\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:35 crc kubenswrapper[4989]: I1006 10:17:35.662684 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:36 crc kubenswrapper[4989]: I1006 10:17:36.105207 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:36 crc kubenswrapper[4989]: I1006 10:17:36.387651 4989 generic.go:334] "Generic (PLEG): container finished" podID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerID="f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba" exitCode=0 Oct 06 10:17:36 crc kubenswrapper[4989]: I1006 10:17:36.387735 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerDied","Data":"f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba"} Oct 06 10:17:36 crc kubenswrapper[4989]: I1006 10:17:36.387792 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerStarted","Data":"b4fd6f663314034d9115cee5ebf9432d2834c4420b4968db312571b905239acd"} Oct 06 10:17:37 crc kubenswrapper[4989]: I1006 10:17:37.402884 4989 generic.go:334] "Generic (PLEG): container finished" podID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerID="6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51" exitCode=0 Oct 06 10:17:37 crc kubenswrapper[4989]: I1006 10:17:37.402993 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerDied","Data":"6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51"} Oct 06 10:17:38 crc kubenswrapper[4989]: I1006 10:17:38.412535 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerStarted","Data":"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5"} Oct 06 10:17:38 crc kubenswrapper[4989]: I1006 10:17:38.434097 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7fqcn" podStartSLOduration=1.876882172 podStartE2EDuration="3.434081481s" podCreationTimestamp="2025-10-06 10:17:35 +0000 UTC" firstStartedPulling="2025-10-06 10:17:36.394511537 +0000 UTC m=+5907.184537117" lastFinishedPulling="2025-10-06 10:17:37.951710836 +0000 UTC m=+5908.741736426" observedRunningTime="2025-10-06 10:17:38.433932057 +0000 UTC m=+5909.223957657" watchObservedRunningTime="2025-10-06 10:17:38.434081481 +0000 UTC m=+5909.224107061" Oct 06 10:17:40 crc kubenswrapper[4989]: I1006 10:17:40.058269 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ctp2p"] Oct 06 10:17:40 crc kubenswrapper[4989]: I1006 10:17:40.066732 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ctp2p"] Oct 06 10:17:41 crc kubenswrapper[4989]: I1006 10:17:41.967785 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f40369-b4d3-46c1-84e2-21d7e34b0d7f" path="/var/lib/kubelet/pods/26f40369-b4d3-46c1-84e2-21d7e34b0d7f/volumes" Oct 06 10:17:45 crc kubenswrapper[4989]: I1006 10:17:45.663379 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:45 crc kubenswrapper[4989]: I1006 10:17:45.664189 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:45 crc kubenswrapper[4989]: I1006 10:17:45.731062 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:46 crc kubenswrapper[4989]: I1006 10:17:46.568634 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:46 crc kubenswrapper[4989]: I1006 10:17:46.634092 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:48 crc kubenswrapper[4989]: I1006 10:17:48.535987 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7fqcn" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="registry-server" containerID="cri-o://ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5" gracePeriod=2 Oct 06 10:17:48 crc kubenswrapper[4989]: I1006 10:17:48.982491 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.112779 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities\") pod \"4ef3f787-727a-47e2-a315-5a741c7900ae\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.112849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content\") pod \"4ef3f787-727a-47e2-a315-5a741c7900ae\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.112876 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwpzb\" (UniqueName: \"kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb\") pod \"4ef3f787-727a-47e2-a315-5a741c7900ae\" (UID: \"4ef3f787-727a-47e2-a315-5a741c7900ae\") " Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.114101 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities" (OuterVolumeSpecName: "utilities") pod "4ef3f787-727a-47e2-a315-5a741c7900ae" (UID: "4ef3f787-727a-47e2-a315-5a741c7900ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.118629 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb" (OuterVolumeSpecName: "kube-api-access-xwpzb") pod "4ef3f787-727a-47e2-a315-5a741c7900ae" (UID: "4ef3f787-727a-47e2-a315-5a741c7900ae"). InnerVolumeSpecName "kube-api-access-xwpzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.127831 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ef3f787-727a-47e2-a315-5a741c7900ae" (UID: "4ef3f787-727a-47e2-a315-5a741c7900ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.214493 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.214519 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ef3f787-727a-47e2-a315-5a741c7900ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.214528 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwpzb\" (UniqueName: \"kubernetes.io/projected/4ef3f787-727a-47e2-a315-5a741c7900ae-kube-api-access-xwpzb\") on node \"crc\" DevicePath \"\"" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.557170 4989 generic.go:334] "Generic (PLEG): container finished" podID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerID="ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5" exitCode=0 Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.557237 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerDied","Data":"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5"} Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.557292 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7fqcn" event={"ID":"4ef3f787-727a-47e2-a315-5a741c7900ae","Type":"ContainerDied","Data":"b4fd6f663314034d9115cee5ebf9432d2834c4420b4968db312571b905239acd"} Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.557302 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7fqcn" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.557360 4989 scope.go:117] "RemoveContainer" containerID="ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.599265 4989 scope.go:117] "RemoveContainer" containerID="6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.634048 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.652890 4989 scope.go:117] "RemoveContainer" containerID="f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.655240 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7fqcn"] Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.699805 4989 scope.go:117] "RemoveContainer" containerID="ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5" Oct 06 10:17:49 crc kubenswrapper[4989]: E1006 10:17:49.700438 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5\": container with ID starting with ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5 not found: ID does not exist" containerID="ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.700485 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5"} err="failed to get container status \"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5\": rpc error: code = NotFound desc = could not find container \"ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5\": container with ID starting with ed92129160c45b32a6bcba16110fce3871b0442fc57aaf87e83549169e96f3a5 not found: ID does not exist" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.700521 4989 scope.go:117] "RemoveContainer" containerID="6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51" Oct 06 10:17:49 crc kubenswrapper[4989]: E1006 10:17:49.701398 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51\": container with ID starting with 6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51 not found: ID does not exist" containerID="6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.701493 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51"} err="failed to get container status \"6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51\": rpc error: code = NotFound desc = could not find container \"6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51\": container with ID starting with 6e5a22dd34eeb6bcff5fb57d187ca697a41109920e140d528ca28a44f5ca1a51 not found: ID does not exist" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.701550 4989 scope.go:117] "RemoveContainer" containerID="f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba" Oct 06 10:17:49 crc kubenswrapper[4989]: E1006 10:17:49.702007 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba\": container with ID starting with f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba not found: ID does not exist" containerID="f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.702053 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba"} err="failed to get container status \"f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba\": rpc error: code = NotFound desc = could not find container \"f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba\": container with ID starting with f909d4d72786e4a298ffc99b8fbd7931032e4954f5372520c0c21537cdbf8dba not found: ID does not exist" Oct 06 10:17:49 crc kubenswrapper[4989]: I1006 10:17:49.959159 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" path="/var/lib/kubelet/pods/4ef3f787-727a-47e2-a315-5a741c7900ae/volumes" Oct 06 10:17:50 crc kubenswrapper[4989]: I1006 10:17:50.034975 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8171-account-create-xvtbw"] Oct 06 10:17:50 crc kubenswrapper[4989]: I1006 10:17:50.066918 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8171-account-create-xvtbw"] Oct 06 10:17:51 crc kubenswrapper[4989]: I1006 10:17:51.950499 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb724a6-b54c-42ae-bcf5-a6fae28d19c5" path="/var/lib/kubelet/pods/1eb724a6-b54c-42ae-bcf5-a6fae28d19c5/volumes" Oct 06 10:17:57 crc kubenswrapper[4989]: I1006 10:17:57.033265 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jfp5d"] Oct 06 10:17:57 crc kubenswrapper[4989]: I1006 10:17:57.047479 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jfp5d"] Oct 06 10:17:57 crc kubenswrapper[4989]: I1006 10:17:57.952258 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f732e7-a08f-41b9-80fc-bcf60cb142d8" path="/var/lib/kubelet/pods/91f732e7-a08f-41b9-80fc-bcf60cb142d8/volumes" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.900358 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:04 crc kubenswrapper[4989]: E1006 10:18:04.903836 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="registry-server" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.904016 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="registry-server" Oct 06 10:18:04 crc kubenswrapper[4989]: E1006 10:18:04.904221 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="extract-utilities" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.904398 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="extract-utilities" Oct 06 10:18:04 crc kubenswrapper[4989]: E1006 10:18:04.904634 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="extract-content" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.904869 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="extract-content" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.905546 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef3f787-727a-47e2-a315-5a741c7900ae" containerName="registry-server" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.908625 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.920960 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.946126 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.946465 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7h7\" (UniqueName: \"kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:04 crc kubenswrapper[4989]: I1006 10:18:04.946530 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.048877 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7h7\" (UniqueName: \"kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.048931 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.049018 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.049633 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.049686 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.072589 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7h7\" (UniqueName: \"kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7\") pod \"community-operators-b8j7j\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.251127 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:05 crc kubenswrapper[4989]: I1006 10:18:05.794568 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:06 crc kubenswrapper[4989]: I1006 10:18:06.738429 4989 generic.go:334] "Generic (PLEG): container finished" podID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerID="20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2" exitCode=0 Oct 06 10:18:06 crc kubenswrapper[4989]: I1006 10:18:06.738525 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerDied","Data":"20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2"} Oct 06 10:18:06 crc kubenswrapper[4989]: I1006 10:18:06.739014 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerStarted","Data":"d798f045a3052e9c38030e63dcaf0f586c529342307ebeffa2b694e4f4c9a879"} Oct 06 10:18:06 crc kubenswrapper[4989]: I1006 10:18:06.741690 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:18:07 crc kubenswrapper[4989]: I1006 10:18:07.766088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerStarted","Data":"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d"} Oct 06 10:18:08 crc kubenswrapper[4989]: I1006 10:18:08.785222 4989 generic.go:334] "Generic (PLEG): container finished" podID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerID="057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d" exitCode=0 Oct 06 10:18:08 crc kubenswrapper[4989]: I1006 10:18:08.785279 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerDied","Data":"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d"} Oct 06 10:18:09 crc kubenswrapper[4989]: I1006 10:18:09.801324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerStarted","Data":"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d"} Oct 06 10:18:09 crc kubenswrapper[4989]: I1006 10:18:09.832034 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b8j7j" podStartSLOduration=3.36758209 podStartE2EDuration="5.832010601s" podCreationTimestamp="2025-10-06 10:18:04 +0000 UTC" firstStartedPulling="2025-10-06 10:18:06.741463601 +0000 UTC m=+5937.531489191" lastFinishedPulling="2025-10-06 10:18:09.205892122 +0000 UTC m=+5939.995917702" observedRunningTime="2025-10-06 10:18:09.819690986 +0000 UTC m=+5940.609716586" watchObservedRunningTime="2025-10-06 10:18:09.832010601 +0000 UTC m=+5940.622036191" Oct 06 10:18:11 crc kubenswrapper[4989]: I1006 10:18:11.055464 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rlv82"] Oct 06 10:18:11 crc kubenswrapper[4989]: I1006 10:18:11.064670 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rlv82"] Oct 06 10:18:11 crc kubenswrapper[4989]: I1006 10:18:11.961510 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adee3030-b6d6-4ba9-9284-c5b62de44d46" path="/var/lib/kubelet/pods/adee3030-b6d6-4ba9-9284-c5b62de44d46/volumes" Oct 06 10:18:15 crc kubenswrapper[4989]: I1006 10:18:15.253162 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:15 crc kubenswrapper[4989]: I1006 10:18:15.253739 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:15 crc kubenswrapper[4989]: I1006 10:18:15.338803 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:15 crc kubenswrapper[4989]: I1006 10:18:15.952345 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:16 crc kubenswrapper[4989]: I1006 10:18:16.001975 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:17 crc kubenswrapper[4989]: I1006 10:18:17.913215 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b8j7j" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="registry-server" containerID="cri-o://2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d" gracePeriod=2 Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.408570 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.472724 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn7h7\" (UniqueName: \"kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7\") pod \"45da8e22-9ac7-470f-a970-17b24afd1b61\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.472804 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content\") pod \"45da8e22-9ac7-470f-a970-17b24afd1b61\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.472905 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities\") pod \"45da8e22-9ac7-470f-a970-17b24afd1b61\" (UID: \"45da8e22-9ac7-470f-a970-17b24afd1b61\") " Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.474724 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities" (OuterVolumeSpecName: "utilities") pod "45da8e22-9ac7-470f-a970-17b24afd1b61" (UID: "45da8e22-9ac7-470f-a970-17b24afd1b61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.478695 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7" (OuterVolumeSpecName: "kube-api-access-jn7h7") pod "45da8e22-9ac7-470f-a970-17b24afd1b61" (UID: "45da8e22-9ac7-470f-a970-17b24afd1b61"). InnerVolumeSpecName "kube-api-access-jn7h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.559842 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45da8e22-9ac7-470f-a970-17b24afd1b61" (UID: "45da8e22-9ac7-470f-a970-17b24afd1b61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.575607 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn7h7\" (UniqueName: \"kubernetes.io/projected/45da8e22-9ac7-470f-a970-17b24afd1b61-kube-api-access-jn7h7\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.575664 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.575678 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45da8e22-9ac7-470f-a970-17b24afd1b61-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.928568 4989 generic.go:334] "Generic (PLEG): container finished" podID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerID="2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d" exitCode=0 Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.928626 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerDied","Data":"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d"} Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.928732 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7j" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.928757 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7j" event={"ID":"45da8e22-9ac7-470f-a970-17b24afd1b61","Type":"ContainerDied","Data":"d798f045a3052e9c38030e63dcaf0f586c529342307ebeffa2b694e4f4c9a879"} Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.928783 4989 scope.go:117] "RemoveContainer" containerID="2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d" Oct 06 10:18:18 crc kubenswrapper[4989]: I1006 10:18:18.977703 4989 scope.go:117] "RemoveContainer" containerID="057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:18.994806 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.004129 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b8j7j"] Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.020160 4989 scope.go:117] "RemoveContainer" containerID="20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.082588 4989 scope.go:117] "RemoveContainer" containerID="2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d" Oct 06 10:18:19 crc kubenswrapper[4989]: E1006 10:18:19.083124 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d\": container with ID starting with 2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d not found: ID does not exist" containerID="2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.083164 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d"} err="failed to get container status \"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d\": rpc error: code = NotFound desc = could not find container \"2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d\": container with ID starting with 2578c91f50d251bfa344222c25415c3f80180eec994acf4c627a95dd345d343d not found: ID does not exist" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.083189 4989 scope.go:117] "RemoveContainer" containerID="057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d" Oct 06 10:18:19 crc kubenswrapper[4989]: E1006 10:18:19.083498 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d\": container with ID starting with 057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d not found: ID does not exist" containerID="057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.083526 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d"} err="failed to get container status \"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d\": rpc error: code = NotFound desc = could not find container \"057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d\": container with ID starting with 057d26d7ab13b57737c2772e713970ea5755616dd6a4ae356881efa8e478164d not found: ID does not exist" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.083546 4989 scope.go:117] "RemoveContainer" containerID="20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2" Oct 06 10:18:19 crc kubenswrapper[4989]: E1006 10:18:19.083830 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2\": container with ID starting with 20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2 not found: ID does not exist" containerID="20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.083867 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2"} err="failed to get container status \"20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2\": rpc error: code = NotFound desc = could not find container \"20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2\": container with ID starting with 20990b7494dbd008472f9839198941dd588ffd4c62d8cc797f311f74fa7ac9d2 not found: ID does not exist" Oct 06 10:18:19 crc kubenswrapper[4989]: I1006 10:18:19.950198 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" path="/var/lib/kubelet/pods/45da8e22-9ac7-470f-a970-17b24afd1b61/volumes" Oct 06 10:18:21 crc kubenswrapper[4989]: I1006 10:18:21.508793 4989 scope.go:117] "RemoveContainer" containerID="e4ced04d64dc10f837231c9958a4e615f3bdee5a1441218894f0339ead0fde15" Oct 06 10:18:21 crc kubenswrapper[4989]: I1006 10:18:21.577323 4989 scope.go:117] "RemoveContainer" containerID="169fe97811ea870cbd51b9bdbc1d3ed8cf6d74b9fe1a49516f4396dddf7cb082" Oct 06 10:18:21 crc kubenswrapper[4989]: I1006 10:18:21.605990 4989 scope.go:117] "RemoveContainer" containerID="82e3fb0a8e9d221cf3a76861dda7db006629cd32643acd24cc14bde43d65fd41" Oct 06 10:18:21 crc kubenswrapper[4989]: I1006 10:18:21.651618 4989 scope.go:117] "RemoveContainer" containerID="9a527d4b848b349bde90deaa00816cde1b68e159e0d5f9f914a7cf564bbd9584" Oct 06 10:18:33 crc kubenswrapper[4989]: I1006 10:18:33.935611 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:18:33 crc kubenswrapper[4989]: I1006 10:18:33.936350 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:19:03 crc kubenswrapper[4989]: I1006 10:19:03.935241 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:19:03 crc kubenswrapper[4989]: I1006 10:19:03.937293 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.881985 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8r8dw"] Oct 06 10:19:09 crc kubenswrapper[4989]: E1006 10:19:09.883146 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="registry-server" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.883167 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="registry-server" Oct 06 10:19:09 crc kubenswrapper[4989]: E1006 10:19:09.883216 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="extract-content" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.883224 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="extract-content" Oct 06 10:19:09 crc kubenswrapper[4989]: E1006 10:19:09.883245 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="extract-utilities" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.883253 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="extract-utilities" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.883537 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="45da8e22-9ac7-470f-a970-17b24afd1b61" containerName="registry-server" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.884598 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.894807 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8r8dw"] Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.900469 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.901049 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7t2rg" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.906060 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-pkfng"] Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.920919 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pkfng"] Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.921053 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.977760 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-log-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.977865 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-etc-ovs\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.977930 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-lib\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.977965 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978057 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-log\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978091 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/858946e9-0129-43b4-9044-7878d1e0b562-scripts\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978112 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk9m4\" (UniqueName: \"kubernetes.io/projected/47747799-142b-42f8-ad3a-ba483ad279d1-kube-api-access-zk9m4\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47747799-142b-42f8-ad3a-ba483ad279d1-scripts\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-run\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978259 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdr7b\" (UniqueName: \"kubernetes.io/projected/858946e9-0129-43b4-9044-7878d1e0b562-kube-api-access-bdr7b\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:09 crc kubenswrapper[4989]: I1006 10:19:09.978287 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079600 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079734 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-log\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079757 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/858946e9-0129-43b4-9044-7878d1e0b562-scripts\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk9m4\" (UniqueName: \"kubernetes.io/projected/47747799-142b-42f8-ad3a-ba483ad279d1-kube-api-access-zk9m4\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079826 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47747799-142b-42f8-ad3a-ba483ad279d1-scripts\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-run\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079903 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdr7b\" (UniqueName: \"kubernetes.io/projected/858946e9-0129-43b4-9044-7878d1e0b562-kube-api-access-bdr7b\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079975 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-log-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.079989 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080002 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-etc-ovs\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080090 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-etc-ovs\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080118 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-lib\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080164 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-lib\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080206 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-run\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080230 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/858946e9-0129-43b4-9044-7878d1e0b562-var-log\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080468 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-run\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.080531 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/47747799-142b-42f8-ad3a-ba483ad279d1-var-log-ovn\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.083199 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.092780 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47747799-142b-42f8-ad3a-ba483ad279d1-scripts\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.093452 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/858946e9-0129-43b4-9044-7878d1e0b562-scripts\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.099619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdr7b\" (UniqueName: \"kubernetes.io/projected/858946e9-0129-43b4-9044-7878d1e0b562-kube-api-access-bdr7b\") pod \"ovn-controller-ovs-pkfng\" (UID: \"858946e9-0129-43b4-9044-7878d1e0b562\") " pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.099870 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk9m4\" (UniqueName: \"kubernetes.io/projected/47747799-142b-42f8-ad3a-ba483ad279d1-kube-api-access-zk9m4\") pod \"ovn-controller-8r8dw\" (UID: \"47747799-142b-42f8-ad3a-ba483ad279d1\") " pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.215248 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7t2rg" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.224004 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.243073 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:10 crc kubenswrapper[4989]: I1006 10:19:10.717628 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8r8dw"] Oct 06 10:19:10 crc kubenswrapper[4989]: W1006 10:19:10.719473 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47747799_142b_42f8_ad3a_ba483ad279d1.slice/crio-8cd1bbb73bb4c48e364104d1eea996070f7e8fa0a61f4bb5bfda69931b2bbda2 WatchSource:0}: Error finding container 8cd1bbb73bb4c48e364104d1eea996070f7e8fa0a61f4bb5bfda69931b2bbda2: Status 404 returned error can't find the container with id 8cd1bbb73bb4c48e364104d1eea996070f7e8fa0a61f4bb5bfda69931b2bbda2 Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.113379 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-pkfng"] Oct 06 10:19:11 crc kubenswrapper[4989]: W1006 10:19:11.122161 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod858946e9_0129_43b4_9044_7878d1e0b562.slice/crio-7553c1424d6af990c29fe3a30814254dc1fc415241e41da0eb20887604935929 WatchSource:0}: Error finding container 7553c1424d6af990c29fe3a30814254dc1fc415241e41da0eb20887604935929: Status 404 returned error can't find the container with id 7553c1424d6af990c29fe3a30814254dc1fc415241e41da0eb20887604935929 Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.435879 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9lbv7"] Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.438586 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.440677 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9lbv7"] Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.442864 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.508321 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb5b6\" (UniqueName: \"kubernetes.io/projected/764a099f-2698-470d-a6bc-d46435ac4951-kube-api-access-pb5b6\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.508423 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovn-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.508506 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764a099f-2698-470d-a6bc-d46435ac4951-config\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.508530 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovs-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.534157 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pkfng" event={"ID":"858946e9-0129-43b4-9044-7878d1e0b562","Type":"ContainerStarted","Data":"1a8644b2589fd4b994d5005e08c3aa8e0595b72396b2ba8fb1c8995d11e82461"} Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.534432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pkfng" event={"ID":"858946e9-0129-43b4-9044-7878d1e0b562","Type":"ContainerStarted","Data":"7553c1424d6af990c29fe3a30814254dc1fc415241e41da0eb20887604935929"} Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.538388 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8r8dw" event={"ID":"47747799-142b-42f8-ad3a-ba483ad279d1","Type":"ContainerStarted","Data":"c64c440a04fa6401cec458794e0a97723f4c6374cd20a0669ba6b0d422c0eebc"} Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.538604 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8r8dw" event={"ID":"47747799-142b-42f8-ad3a-ba483ad279d1","Type":"ContainerStarted","Data":"8cd1bbb73bb4c48e364104d1eea996070f7e8fa0a61f4bb5bfda69931b2bbda2"} Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.539228 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.571739 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8r8dw" podStartSLOduration=2.5717227019999997 podStartE2EDuration="2.571722702s" podCreationTimestamp="2025-10-06 10:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:11.568314864 +0000 UTC m=+6002.358340434" watchObservedRunningTime="2025-10-06 10:19:11.571722702 +0000 UTC m=+6002.361748282" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.610265 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764a099f-2698-470d-a6bc-d46435ac4951-config\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.610345 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovs-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.610385 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb5b6\" (UniqueName: \"kubernetes.io/projected/764a099f-2698-470d-a6bc-d46435ac4951-kube-api-access-pb5b6\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.610489 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovn-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.610801 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovn-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.611081 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/764a099f-2698-470d-a6bc-d46435ac4951-ovs-rundir\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.611497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/764a099f-2698-470d-a6bc-d46435ac4951-config\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.637188 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb5b6\" (UniqueName: \"kubernetes.io/projected/764a099f-2698-470d-a6bc-d46435ac4951-kube-api-access-pb5b6\") pod \"ovn-controller-metrics-9lbv7\" (UID: \"764a099f-2698-470d-a6bc-d46435ac4951\") " pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:11 crc kubenswrapper[4989]: I1006 10:19:11.774560 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9lbv7" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.233600 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9lbv7"] Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.411591 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-lqqtw"] Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.413434 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.446695 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lqqtw"] Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.525913 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzft4\" (UniqueName: \"kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4\") pod \"octavia-db-create-lqqtw\" (UID: \"987d1d1a-c7a5-4122-a2eb-86b811863dd5\") " pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.551319 4989 generic.go:334] "Generic (PLEG): container finished" podID="858946e9-0129-43b4-9044-7878d1e0b562" containerID="1a8644b2589fd4b994d5005e08c3aa8e0595b72396b2ba8fb1c8995d11e82461" exitCode=0 Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.551371 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pkfng" event={"ID":"858946e9-0129-43b4-9044-7878d1e0b562","Type":"ContainerDied","Data":"1a8644b2589fd4b994d5005e08c3aa8e0595b72396b2ba8fb1c8995d11e82461"} Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.553900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9lbv7" event={"ID":"764a099f-2698-470d-a6bc-d46435ac4951","Type":"ContainerStarted","Data":"dd1bb052d3a054ea18de7dc33e573fa124d46e9d01debabf22ec45123eda1be8"} Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.553942 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9lbv7" event={"ID":"764a099f-2698-470d-a6bc-d46435ac4951","Type":"ContainerStarted","Data":"fcc3fa342f042c1a8c9c419ddfea8d79509a2379f88d52f5b9804a2a6ccc3e57"} Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.587207 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9lbv7" podStartSLOduration=1.587187285 podStartE2EDuration="1.587187285s" podCreationTimestamp="2025-10-06 10:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:12.584720334 +0000 UTC m=+6003.374745934" watchObservedRunningTime="2025-10-06 10:19:12.587187285 +0000 UTC m=+6003.377212865" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.627558 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzft4\" (UniqueName: \"kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4\") pod \"octavia-db-create-lqqtw\" (UID: \"987d1d1a-c7a5-4122-a2eb-86b811863dd5\") " pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.653241 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzft4\" (UniqueName: \"kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4\") pod \"octavia-db-create-lqqtw\" (UID: \"987d1d1a-c7a5-4122-a2eb-86b811863dd5\") " pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:12 crc kubenswrapper[4989]: I1006 10:19:12.738813 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.222551 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-lqqtw"] Oct 06 10:19:13 crc kubenswrapper[4989]: W1006 10:19:13.228531 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod987d1d1a_c7a5_4122_a2eb_86b811863dd5.slice/crio-72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50 WatchSource:0}: Error finding container 72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50: Status 404 returned error can't find the container with id 72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50 Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.566827 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pkfng" event={"ID":"858946e9-0129-43b4-9044-7878d1e0b562","Type":"ContainerStarted","Data":"0e7b2bbe7130903af5c5de4b23a5083d4c97df463dc88f1803cdd50946621540"} Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.567161 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-pkfng" event={"ID":"858946e9-0129-43b4-9044-7878d1e0b562","Type":"ContainerStarted","Data":"c8372b52188ccbd7083bfaa832c9774a2d1a57ed76e20d716fc55ba8e8d91546"} Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.567179 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.567191 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.576055 4989 generic.go:334] "Generic (PLEG): container finished" podID="987d1d1a-c7a5-4122-a2eb-86b811863dd5" containerID="b19d4ca0d4149c76229263d7308475c7e74976d55ecba36b6f416803dc08b9bc" exitCode=0 Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.576130 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lqqtw" event={"ID":"987d1d1a-c7a5-4122-a2eb-86b811863dd5","Type":"ContainerDied","Data":"b19d4ca0d4149c76229263d7308475c7e74976d55ecba36b6f416803dc08b9bc"} Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.576182 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lqqtw" event={"ID":"987d1d1a-c7a5-4122-a2eb-86b811863dd5","Type":"ContainerStarted","Data":"72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50"} Oct 06 10:19:13 crc kubenswrapper[4989]: I1006 10:19:13.593938 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-pkfng" podStartSLOduration=4.593904055 podStartE2EDuration="4.593904055s" podCreationTimestamp="2025-10-06 10:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:19:13.588012735 +0000 UTC m=+6004.378038305" watchObservedRunningTime="2025-10-06 10:19:13.593904055 +0000 UTC m=+6004.383929665" Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.000481 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.090700 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzft4\" (UniqueName: \"kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4\") pod \"987d1d1a-c7a5-4122-a2eb-86b811863dd5\" (UID: \"987d1d1a-c7a5-4122-a2eb-86b811863dd5\") " Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.096489 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4" (OuterVolumeSpecName: "kube-api-access-gzft4") pod "987d1d1a-c7a5-4122-a2eb-86b811863dd5" (UID: "987d1d1a-c7a5-4122-a2eb-86b811863dd5"). InnerVolumeSpecName "kube-api-access-gzft4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.193435 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzft4\" (UniqueName: \"kubernetes.io/projected/987d1d1a-c7a5-4122-a2eb-86b811863dd5-kube-api-access-gzft4\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.608505 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-lqqtw" event={"ID":"987d1d1a-c7a5-4122-a2eb-86b811863dd5","Type":"ContainerDied","Data":"72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50"} Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.608578 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72792a361d88e09b20da20ee42576cbe66321d9c3dd0e0f4fd8c91695c067c50" Oct 06 10:19:15 crc kubenswrapper[4989]: I1006 10:19:15.608597 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-lqqtw" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.727327 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-7f15-account-create-nhcgx"] Oct 06 10:19:24 crc kubenswrapper[4989]: E1006 10:19:24.728235 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987d1d1a-c7a5-4122-a2eb-86b811863dd5" containerName="mariadb-database-create" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.728317 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="987d1d1a-c7a5-4122-a2eb-86b811863dd5" containerName="mariadb-database-create" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.728501 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="987d1d1a-c7a5-4122-a2eb-86b811863dd5" containerName="mariadb-database-create" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.729133 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.730912 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.741258 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-7f15-account-create-nhcgx"] Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.895403 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9fl2\" (UniqueName: \"kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2\") pod \"octavia-7f15-account-create-nhcgx\" (UID: \"ad4080dd-79fe-4dcd-88f9-80ddb79ab629\") " pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:24 crc kubenswrapper[4989]: I1006 10:19:24.998255 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9fl2\" (UniqueName: \"kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2\") pod \"octavia-7f15-account-create-nhcgx\" (UID: \"ad4080dd-79fe-4dcd-88f9-80ddb79ab629\") " pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:25 crc kubenswrapper[4989]: I1006 10:19:25.029541 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9fl2\" (UniqueName: \"kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2\") pod \"octavia-7f15-account-create-nhcgx\" (UID: \"ad4080dd-79fe-4dcd-88f9-80ddb79ab629\") " pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:25 crc kubenswrapper[4989]: I1006 10:19:25.045760 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:25 crc kubenswrapper[4989]: I1006 10:19:25.607674 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-7f15-account-create-nhcgx"] Oct 06 10:19:25 crc kubenswrapper[4989]: I1006 10:19:25.704126 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7f15-account-create-nhcgx" event={"ID":"ad4080dd-79fe-4dcd-88f9-80ddb79ab629","Type":"ContainerStarted","Data":"feb6815442b5ef1f3444ed332e32f2bf6f69876e942cd9b0b8eda703902f714e"} Oct 06 10:19:26 crc kubenswrapper[4989]: I1006 10:19:26.715151 4989 generic.go:334] "Generic (PLEG): container finished" podID="ad4080dd-79fe-4dcd-88f9-80ddb79ab629" containerID="ff0a30952b8f7865acfb4f535750171b9456663c8f65acdcfc46f81cd415a23d" exitCode=0 Oct 06 10:19:26 crc kubenswrapper[4989]: I1006 10:19:26.715197 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7f15-account-create-nhcgx" event={"ID":"ad4080dd-79fe-4dcd-88f9-80ddb79ab629","Type":"ContainerDied","Data":"ff0a30952b8f7865acfb4f535750171b9456663c8f65acdcfc46f81cd415a23d"} Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.089001 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.196141 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9fl2\" (UniqueName: \"kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2\") pod \"ad4080dd-79fe-4dcd-88f9-80ddb79ab629\" (UID: \"ad4080dd-79fe-4dcd-88f9-80ddb79ab629\") " Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.204003 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2" (OuterVolumeSpecName: "kube-api-access-s9fl2") pod "ad4080dd-79fe-4dcd-88f9-80ddb79ab629" (UID: "ad4080dd-79fe-4dcd-88f9-80ddb79ab629"). InnerVolumeSpecName "kube-api-access-s9fl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.300125 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9fl2\" (UniqueName: \"kubernetes.io/projected/ad4080dd-79fe-4dcd-88f9-80ddb79ab629-kube-api-access-s9fl2\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.746013 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-7f15-account-create-nhcgx" event={"ID":"ad4080dd-79fe-4dcd-88f9-80ddb79ab629","Type":"ContainerDied","Data":"feb6815442b5ef1f3444ed332e32f2bf6f69876e942cd9b0b8eda703902f714e"} Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.746087 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feb6815442b5ef1f3444ed332e32f2bf6f69876e942cd9b0b8eda703902f714e" Oct 06 10:19:28 crc kubenswrapper[4989]: I1006 10:19:28.746088 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-7f15-account-create-nhcgx" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.535306 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-kxqsb"] Oct 06 10:19:31 crc kubenswrapper[4989]: E1006 10:19:31.536114 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4080dd-79fe-4dcd-88f9-80ddb79ab629" containerName="mariadb-account-create" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.536130 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4080dd-79fe-4dcd-88f9-80ddb79ab629" containerName="mariadb-account-create" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.536351 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4080dd-79fe-4dcd-88f9-80ddb79ab629" containerName="mariadb-account-create" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.537093 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.554909 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-kxqsb"] Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.672128 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b2rr\" (UniqueName: \"kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr\") pod \"octavia-persistence-db-create-kxqsb\" (UID: \"2e52fd09-c241-4aa6-b94d-5e0a653c7f10\") " pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.773750 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b2rr\" (UniqueName: \"kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr\") pod \"octavia-persistence-db-create-kxqsb\" (UID: \"2e52fd09-c241-4aa6-b94d-5e0a653c7f10\") " pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.801308 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b2rr\" (UniqueName: \"kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr\") pod \"octavia-persistence-db-create-kxqsb\" (UID: \"2e52fd09-c241-4aa6-b94d-5e0a653c7f10\") " pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:31 crc kubenswrapper[4989]: I1006 10:19:31.860293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:32 crc kubenswrapper[4989]: I1006 10:19:32.316893 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-kxqsb"] Oct 06 10:19:32 crc kubenswrapper[4989]: W1006 10:19:32.328836 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e52fd09_c241_4aa6_b94d_5e0a653c7f10.slice/crio-4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1 WatchSource:0}: Error finding container 4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1: Status 404 returned error can't find the container with id 4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1 Oct 06 10:19:32 crc kubenswrapper[4989]: I1006 10:19:32.806924 4989 generic.go:334] "Generic (PLEG): container finished" podID="2e52fd09-c241-4aa6-b94d-5e0a653c7f10" containerID="1c8a5cff745212002f5fe26f49aacad0daa715a38f7f7d9956b532869cd6da74" exitCode=0 Oct 06 10:19:32 crc kubenswrapper[4989]: I1006 10:19:32.807005 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-kxqsb" event={"ID":"2e52fd09-c241-4aa6-b94d-5e0a653c7f10","Type":"ContainerDied","Data":"1c8a5cff745212002f5fe26f49aacad0daa715a38f7f7d9956b532869cd6da74"} Oct 06 10:19:32 crc kubenswrapper[4989]: I1006 10:19:32.807241 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-kxqsb" event={"ID":"2e52fd09-c241-4aa6-b94d-5e0a653c7f10","Type":"ContainerStarted","Data":"4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1"} Oct 06 10:19:33 crc kubenswrapper[4989]: I1006 10:19:33.935421 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:19:33 crc kubenswrapper[4989]: I1006 10:19:33.935494 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:19:33 crc kubenswrapper[4989]: I1006 10:19:33.951511 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:19:33 crc kubenswrapper[4989]: I1006 10:19:33.952226 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:19:33 crc kubenswrapper[4989]: I1006 10:19:33.952282 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" gracePeriod=600 Oct 06 10:19:34 crc kubenswrapper[4989]: E1006 10:19:34.096124 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.146309 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.327855 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b2rr\" (UniqueName: \"kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr\") pod \"2e52fd09-c241-4aa6-b94d-5e0a653c7f10\" (UID: \"2e52fd09-c241-4aa6-b94d-5e0a653c7f10\") " Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.333901 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr" (OuterVolumeSpecName: "kube-api-access-2b2rr") pod "2e52fd09-c241-4aa6-b94d-5e0a653c7f10" (UID: "2e52fd09-c241-4aa6-b94d-5e0a653c7f10"). InnerVolumeSpecName "kube-api-access-2b2rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.431283 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b2rr\" (UniqueName: \"kubernetes.io/projected/2e52fd09-c241-4aa6-b94d-5e0a653c7f10-kube-api-access-2b2rr\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.827133 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-kxqsb" event={"ID":"2e52fd09-c241-4aa6-b94d-5e0a653c7f10","Type":"ContainerDied","Data":"4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1"} Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.827192 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f8e15f15cf48572f82dc97d38d092c59381c6f6752cbc3de2b009778f08eaa1" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.827194 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-kxqsb" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.830221 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" exitCode=0 Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.830278 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb"} Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.830336 4989 scope.go:117] "RemoveContainer" containerID="ca408d738ea5ba1a35a8ddc737f7d43078b799668bcd2236905b660c800c50e0" Oct 06 10:19:34 crc kubenswrapper[4989]: I1006 10:19:34.831948 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:19:34 crc kubenswrapper[4989]: E1006 10:19:34.832793 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.681556 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-4547-account-create-wz8wn"] Oct 06 10:19:42 crc kubenswrapper[4989]: E1006 10:19:42.684921 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e52fd09-c241-4aa6-b94d-5e0a653c7f10" containerName="mariadb-database-create" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.685125 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e52fd09-c241-4aa6-b94d-5e0a653c7f10" containerName="mariadb-database-create" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.685807 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e52fd09-c241-4aa6-b94d-5e0a653c7f10" containerName="mariadb-database-create" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.687502 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.690370 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-4547-account-create-wz8wn"] Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.691271 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.817136 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm2qs\" (UniqueName: \"kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs\") pod \"octavia-4547-account-create-wz8wn\" (UID: \"1513a90e-5b01-4b75-a007-97aaf1daa3d7\") " pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.919297 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm2qs\" (UniqueName: \"kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs\") pod \"octavia-4547-account-create-wz8wn\" (UID: \"1513a90e-5b01-4b75-a007-97aaf1daa3d7\") " pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:42 crc kubenswrapper[4989]: I1006 10:19:42.960605 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm2qs\" (UniqueName: \"kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs\") pod \"octavia-4547-account-create-wz8wn\" (UID: \"1513a90e-5b01-4b75-a007-97aaf1daa3d7\") " pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:43 crc kubenswrapper[4989]: I1006 10:19:43.028386 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:43 crc kubenswrapper[4989]: I1006 10:19:43.511515 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-4547-account-create-wz8wn"] Oct 06 10:19:43 crc kubenswrapper[4989]: I1006 10:19:43.949502 4989 generic.go:334] "Generic (PLEG): container finished" podID="1513a90e-5b01-4b75-a007-97aaf1daa3d7" containerID="dcbbe13bc2e4f01bdd8a95e7d264a2c870dbed5f1fcf2f39737146096d6eb6cf" exitCode=0 Oct 06 10:19:43 crc kubenswrapper[4989]: I1006 10:19:43.952791 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-4547-account-create-wz8wn" event={"ID":"1513a90e-5b01-4b75-a007-97aaf1daa3d7","Type":"ContainerDied","Data":"dcbbe13bc2e4f01bdd8a95e7d264a2c870dbed5f1fcf2f39737146096d6eb6cf"} Oct 06 10:19:43 crc kubenswrapper[4989]: I1006 10:19:43.952853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-4547-account-create-wz8wn" event={"ID":"1513a90e-5b01-4b75-a007-97aaf1daa3d7","Type":"ContainerStarted","Data":"628047f624827a46e420db406f0900563eae1585e4e5528c3aca480b59948f35"} Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.280288 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8r8dw" podUID="47747799-142b-42f8-ad3a-ba483ad279d1" containerName="ovn-controller" probeResult="failure" output=< Oct 06 10:19:45 crc kubenswrapper[4989]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 06 10:19:45 crc kubenswrapper[4989]: > Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.294534 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.327421 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-pkfng" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.378522 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.449533 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8r8dw-config-q8bpq"] Oct 06 10:19:45 crc kubenswrapper[4989]: E1006 10:19:45.450214 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1513a90e-5b01-4b75-a007-97aaf1daa3d7" containerName="mariadb-account-create" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.450248 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1513a90e-5b01-4b75-a007-97aaf1daa3d7" containerName="mariadb-account-create" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.450621 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1513a90e-5b01-4b75-a007-97aaf1daa3d7" containerName="mariadb-account-create" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.451623 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.456596 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.464931 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8r8dw-config-q8bpq"] Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.477269 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm2qs\" (UniqueName: \"kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs\") pod \"1513a90e-5b01-4b75-a007-97aaf1daa3d7\" (UID: \"1513a90e-5b01-4b75-a007-97aaf1daa3d7\") " Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.490928 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs" (OuterVolumeSpecName: "kube-api-access-gm2qs") pod "1513a90e-5b01-4b75-a007-97aaf1daa3d7" (UID: "1513a90e-5b01-4b75-a007-97aaf1daa3d7"). InnerVolumeSpecName "kube-api-access-gm2qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.579448 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.579497 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.579835 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.579889 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.579946 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5k9q\" (UniqueName: \"kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.580118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.580202 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm2qs\" (UniqueName: \"kubernetes.io/projected/1513a90e-5b01-4b75-a007-97aaf1daa3d7-kube-api-access-gm2qs\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.681367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5k9q\" (UniqueName: \"kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.681746 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682012 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682074 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682104 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682902 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682252 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.682791 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.683214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.684692 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.711436 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5k9q\" (UniqueName: \"kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q\") pod \"ovn-controller-8r8dw-config-q8bpq\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.775016 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.937835 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:19:45 crc kubenswrapper[4989]: E1006 10:19:45.938132 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.973808 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-4547-account-create-wz8wn" Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.973856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-4547-account-create-wz8wn" event={"ID":"1513a90e-5b01-4b75-a007-97aaf1daa3d7","Type":"ContainerDied","Data":"628047f624827a46e420db406f0900563eae1585e4e5528c3aca480b59948f35"} Oct 06 10:19:45 crc kubenswrapper[4989]: I1006 10:19:45.973883 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628047f624827a46e420db406f0900563eae1585e4e5528c3aca480b59948f35" Oct 06 10:19:46 crc kubenswrapper[4989]: I1006 10:19:46.246107 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8r8dw-config-q8bpq"] Oct 06 10:19:46 crc kubenswrapper[4989]: W1006 10:19:46.256023 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39db4b19_dc19_4975_a72e_9959a747df2e.slice/crio-82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf WatchSource:0}: Error finding container 82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf: Status 404 returned error can't find the container with id 82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf Oct 06 10:19:46 crc kubenswrapper[4989]: I1006 10:19:46.988370 4989 generic.go:334] "Generic (PLEG): container finished" podID="39db4b19-dc19-4975-a72e-9959a747df2e" containerID="bc5254157387b769a9aea3536383115a1d98305236839efbbdcfd48b8a7170f9" exitCode=0 Oct 06 10:19:46 crc kubenswrapper[4989]: I1006 10:19:46.988426 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8r8dw-config-q8bpq" event={"ID":"39db4b19-dc19-4975-a72e-9959a747df2e","Type":"ContainerDied","Data":"bc5254157387b769a9aea3536383115a1d98305236839efbbdcfd48b8a7170f9"} Oct 06 10:19:46 crc kubenswrapper[4989]: I1006 10:19:46.988746 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8r8dw-config-q8bpq" event={"ID":"39db4b19-dc19-4975-a72e-9959a747df2e","Type":"ContainerStarted","Data":"82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf"} Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.447193 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540302 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540486 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5k9q\" (UniqueName: \"kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540577 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540646 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540819 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.540988 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn\") pod \"39db4b19-dc19-4975-a72e-9959a747df2e\" (UID: \"39db4b19-dc19-4975-a72e-9959a747df2e\") " Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.541924 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.542808 4989 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.542909 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.543017 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run" (OuterVolumeSpecName: "var-run") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.544107 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts" (OuterVolumeSpecName: "scripts") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.545710 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.547565 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q" (OuterVolumeSpecName: "kube-api-access-c5k9q") pod "39db4b19-dc19-4975-a72e-9959a747df2e" (UID: "39db4b19-dc19-4975-a72e-9959a747df2e"). InnerVolumeSpecName "kube-api-access-c5k9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.644972 4989 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.645198 4989 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39db4b19-dc19-4975-a72e-9959a747df2e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.645266 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5k9q\" (UniqueName: \"kubernetes.io/projected/39db4b19-dc19-4975-a72e-9959a747df2e-kube-api-access-c5k9q\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.645329 4989 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:48 crc kubenswrapper[4989]: I1006 10:19:48.645383 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39db4b19-dc19-4975-a72e-9959a747df2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.013076 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8r8dw-config-q8bpq" event={"ID":"39db4b19-dc19-4975-a72e-9959a747df2e","Type":"ContainerDied","Data":"82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf"} Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.013128 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82c2bc5a11f8b57ab0134a55aebaa53a76536faf7b9109e6cc2778b97ad367cf" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.013127 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8r8dw-config-q8bpq" Oct 06 10:19:49 crc kubenswrapper[4989]: E1006 10:19:49.140505 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39db4b19_dc19_4975_a72e_9959a747df2e.slice\": RecentStats: unable to find data in memory cache]" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.539532 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8r8dw-config-q8bpq"] Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.549015 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8r8dw-config-q8bpq"] Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.933575 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-db77676cd-jw67c"] Oct 06 10:19:49 crc kubenswrapper[4989]: E1006 10:19:49.934196 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39db4b19-dc19-4975-a72e-9959a747df2e" containerName="ovn-config" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.934223 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="39db4b19-dc19-4975-a72e-9959a747df2e" containerName="ovn-config" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.934503 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="39db4b19-dc19-4975-a72e-9959a747df2e" containerName="ovn-config" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.943492 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.948063 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-t5bvv" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.948263 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.948279 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.960032 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39db4b19-dc19-4975-a72e-9959a747df2e" path="/var/lib/kubelet/pods/39db4b19-dc19-4975-a72e-9959a747df2e/volumes" Oct 06 10:19:49 crc kubenswrapper[4989]: I1006 10:19:49.960839 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-db77676cd-jw67c"] Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.073392 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-scripts\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.073490 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data-merged\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.073584 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-octavia-run\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.073632 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.073677 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-combined-ca-bundle\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.175453 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.175801 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-combined-ca-bundle\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.176599 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-scripts\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.176678 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data-merged\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.176776 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-octavia-run\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.177260 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-octavia-run\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.177613 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data-merged\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.181121 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-config-data\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.181218 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-combined-ca-bundle\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.187700 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfbd9202-a3ec-478c-a15d-40927aa5cf9a-scripts\") pod \"octavia-api-db77676cd-jw67c\" (UID: \"dfbd9202-a3ec-478c-a15d-40927aa5cf9a\") " pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.272559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.278325 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8r8dw" Oct 06 10:19:50 crc kubenswrapper[4989]: I1006 10:19:50.763386 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-db77676cd-jw67c"] Oct 06 10:19:51 crc kubenswrapper[4989]: I1006 10:19:51.045105 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db77676cd-jw67c" event={"ID":"dfbd9202-a3ec-478c-a15d-40927aa5cf9a","Type":"ContainerStarted","Data":"7124894aed4227ae781e3e3526871b785d0a9c7639482e22598c22c2858f4d09"} Oct 06 10:19:57 crc kubenswrapper[4989]: I1006 10:19:57.937238 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:19:57 crc kubenswrapper[4989]: E1006 10:19:57.937974 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:19:59 crc kubenswrapper[4989]: I1006 10:19:59.139704 4989 generic.go:334] "Generic (PLEG): container finished" podID="dfbd9202-a3ec-478c-a15d-40927aa5cf9a" containerID="abcf209a2a05ebb89e123ceab56e592ac64d5725c80238f735252322c6ba990c" exitCode=0 Oct 06 10:19:59 crc kubenswrapper[4989]: I1006 10:19:59.139771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db77676cd-jw67c" event={"ID":"dfbd9202-a3ec-478c-a15d-40927aa5cf9a","Type":"ContainerDied","Data":"abcf209a2a05ebb89e123ceab56e592ac64d5725c80238f735252322c6ba990c"} Oct 06 10:20:00 crc kubenswrapper[4989]: I1006 10:20:00.153586 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db77676cd-jw67c" event={"ID":"dfbd9202-a3ec-478c-a15d-40927aa5cf9a","Type":"ContainerStarted","Data":"c1e58d6a5c8a4d29dcfb8469566b15a98bbb313ec0865ce03479c0db4412b05b"} Oct 06 10:20:00 crc kubenswrapper[4989]: I1006 10:20:00.154130 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-db77676cd-jw67c" event={"ID":"dfbd9202-a3ec-478c-a15d-40927aa5cf9a","Type":"ContainerStarted","Data":"6db0e091f9451d8213ca4fca28a5ee59663ff7170c66e221dbb7683a23c2f732"} Oct 06 10:20:00 crc kubenswrapper[4989]: I1006 10:20:00.154193 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:20:00 crc kubenswrapper[4989]: I1006 10:20:00.154222 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:20:00 crc kubenswrapper[4989]: I1006 10:20:00.188060 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-db77676cd-jw67c" podStartSLOduration=3.317298393 podStartE2EDuration="11.188040341s" podCreationTimestamp="2025-10-06 10:19:49 +0000 UTC" firstStartedPulling="2025-10-06 10:19:50.784355914 +0000 UTC m=+6041.574381494" lastFinishedPulling="2025-10-06 10:19:58.655097862 +0000 UTC m=+6049.445123442" observedRunningTime="2025-10-06 10:20:00.175920772 +0000 UTC m=+6050.965946372" watchObservedRunningTime="2025-10-06 10:20:00.188040341 +0000 UTC m=+6050.978065921" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.324407 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.353129 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-db77676cd-jw67c" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.872352 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-wsxs8"] Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.877113 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.880240 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.880243 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.880714 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.890897 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-wsxs8"] Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.977211 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9294460e-ff38-4149-b42f-5ca293adb874-hm-ports\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.977317 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-scripts\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.977401 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-config-data\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:09 crc kubenswrapper[4989]: I1006 10:20:09.977432 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9294460e-ff38-4149-b42f-5ca293adb874-config-data-merged\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.080207 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9294460e-ff38-4149-b42f-5ca293adb874-hm-ports\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.080633 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-scripts\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.080865 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-config-data\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.081007 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9294460e-ff38-4149-b42f-5ca293adb874-config-data-merged\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.081384 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9294460e-ff38-4149-b42f-5ca293adb874-config-data-merged\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.082167 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.082383 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.082453 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.091952 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9294460e-ff38-4149-b42f-5ca293adb874-hm-ports\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.097560 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-config-data\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.099429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9294460e-ff38-4149-b42f-5ca293adb874-scripts\") pod \"octavia-rsyslog-wsxs8\" (UID: \"9294460e-ff38-4149-b42f-5ca293adb874\") " pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.220683 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.776381 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-wsxs8"] Oct 06 10:20:10 crc kubenswrapper[4989]: W1006 10:20:10.781753 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9294460e_ff38_4149_b42f_5ca293adb874.slice/crio-56d6be24ccb3e85e44416532d5a7b6c770beeef3c1f90bc844b9309a7543f274 WatchSource:0}: Error finding container 56d6be24ccb3e85e44416532d5a7b6c770beeef3c1f90bc844b9309a7543f274: Status 404 returned error can't find the container with id 56d6be24ccb3e85e44416532d5a7b6c770beeef3c1f90bc844b9309a7543f274 Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.889003 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.892536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.894923 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 06 10:20:10 crc kubenswrapper[4989]: I1006 10:20:10.902533 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.000709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.000783 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.102431 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.102945 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.103419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.108485 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config\") pod \"octavia-image-upload-59f8cff499-gfppm\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.232625 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.292639 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-wsxs8" event={"ID":"9294460e-ff38-4149-b42f-5ca293adb874","Type":"ContainerStarted","Data":"56d6be24ccb3e85e44416532d5a7b6c770beeef3c1f90bc844b9309a7543f274"} Oct 06 10:20:11 crc kubenswrapper[4989]: I1006 10:20:11.704617 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:12 crc kubenswrapper[4989]: I1006 10:20:12.305038 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gfppm" event={"ID":"fe4a4862-c2c3-4750-ad47-94dee1b0130b","Type":"ContainerStarted","Data":"204b1a5ae17c51afa395e88e1cbe94b92cd1473f7aec4a7a77b8114bf5b371f6"} Oct 06 10:20:12 crc kubenswrapper[4989]: I1006 10:20:12.938014 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:20:12 crc kubenswrapper[4989]: E1006 10:20:12.938311 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:20:13 crc kubenswrapper[4989]: I1006 10:20:13.319857 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-wsxs8" event={"ID":"9294460e-ff38-4149-b42f-5ca293adb874","Type":"ContainerStarted","Data":"41de3a1a62f72718c1223847f193a1bc3b979e004b699c095e268c80f0d41df3"} Oct 06 10:20:15 crc kubenswrapper[4989]: I1006 10:20:15.339005 4989 generic.go:334] "Generic (PLEG): container finished" podID="9294460e-ff38-4149-b42f-5ca293adb874" containerID="41de3a1a62f72718c1223847f193a1bc3b979e004b699c095e268c80f0d41df3" exitCode=0 Oct 06 10:20:15 crc kubenswrapper[4989]: I1006 10:20:15.339107 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-wsxs8" event={"ID":"9294460e-ff38-4149-b42f-5ca293adb874","Type":"ContainerDied","Data":"41de3a1a62f72718c1223847f193a1bc3b979e004b699c095e268c80f0d41df3"} Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.209978 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-p2zxx"] Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.212057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.214379 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.214577 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.214676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.228339 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-p2zxx"] Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.370599 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-wsxs8" event={"ID":"9294460e-ff38-4149-b42f-5ca293adb874","Type":"ContainerStarted","Data":"f97c5fe797a281d51a699c912cf953900c60d4e2ad8a94be8734570e448ae770"} Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372692 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-amphora-certs\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372749 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372779 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data-merged\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372836 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-combined-ca-bundle\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372870 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-scripts\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.372900 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-hm-ports\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.375662 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.398520 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-wsxs8" podStartSLOduration=2.808752959 podStartE2EDuration="8.398501144s" podCreationTimestamp="2025-10-06 10:20:09 +0000 UTC" firstStartedPulling="2025-10-06 10:20:10.784690838 +0000 UTC m=+6061.574716408" lastFinishedPulling="2025-10-06 10:20:16.374439013 +0000 UTC m=+6067.164464593" observedRunningTime="2025-10-06 10:20:17.393560191 +0000 UTC m=+6068.183585791" watchObservedRunningTime="2025-10-06 10:20:17.398501144 +0000 UTC m=+6068.188526714" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-amphora-certs\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475695 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475735 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data-merged\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475806 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-combined-ca-bundle\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475844 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-scripts\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.475881 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-hm-ports\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.477857 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data-merged\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.478423 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-hm-ports\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.482883 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-combined-ca-bundle\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.483986 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-amphora-certs\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.487528 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-scripts\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.497029 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3e3fae2-2565-4d6b-8a43-30e48c1f734c-config-data\") pod \"octavia-healthmanager-p2zxx\" (UID: \"b3e3fae2-2565-4d6b-8a43-30e48c1f734c\") " pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:17 crc kubenswrapper[4989]: I1006 10:20:17.528291 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.149772 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-p2zxx"] Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.646782 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-jcn7j"] Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.649332 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.651640 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.666840 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-jcn7j"] Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.802402 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.802453 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.802687 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.803160 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.904745 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.904952 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.904994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.905014 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.905543 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.911460 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.911761 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.912617 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts\") pod \"octavia-db-sync-jcn7j\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:18 crc kubenswrapper[4989]: I1006 10:20:18.975572 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.390646 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-7j2m4"] Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.392600 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.395874 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.396023 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.401437 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-7j2m4"] Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536254 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a2fcb796-037b-403e-a59b-2d73a6e94de7-hm-ports\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536335 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data-merged\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536371 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-amphora-certs\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536386 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-scripts\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536507 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-combined-ca-bundle\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.536560 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638145 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-combined-ca-bundle\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638199 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638274 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a2fcb796-037b-403e-a59b-2d73a6e94de7-hm-ports\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638489 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data-merged\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638507 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-amphora-certs\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.638530 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-scripts\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.639242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a2fcb796-037b-403e-a59b-2d73a6e94de7-hm-ports\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.641504 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data-merged\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.644179 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-amphora-certs\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.645377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-config-data\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.646140 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-scripts\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.654599 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fcb796-037b-403e-a59b-2d73a6e94de7-combined-ca-bundle\") pod \"octavia-housekeeping-7j2m4\" (UID: \"a2fcb796-037b-403e-a59b-2d73a6e94de7\") " pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: I1006 10:20:20.722043 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:20 crc kubenswrapper[4989]: W1006 10:20:20.917949 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3e3fae2_2565_4d6b_8a43_30e48c1f734c.slice/crio-3a7c9490851f90a2042f8b1bf4fbc4b56494de1dc6c0ce6f59105eee6e738b1d WatchSource:0}: Error finding container 3a7c9490851f90a2042f8b1bf4fbc4b56494de1dc6c0ce6f59105eee6e738b1d: Status 404 returned error can't find the container with id 3a7c9490851f90a2042f8b1bf4fbc4b56494de1dc6c0ce6f59105eee6e738b1d Oct 06 10:20:21 crc kubenswrapper[4989]: I1006 10:20:21.415592 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-p2zxx" event={"ID":"b3e3fae2-2565-4d6b-8a43-30e48c1f734c","Type":"ContainerStarted","Data":"3a7c9490851f90a2042f8b1bf4fbc4b56494de1dc6c0ce6f59105eee6e738b1d"} Oct 06 10:20:21 crc kubenswrapper[4989]: I1006 10:20:21.543303 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-jcn7j"] Oct 06 10:20:21 crc kubenswrapper[4989]: W1006 10:20:21.551099 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46711b9f_dd60_4901_9f45_a3301b4ea9c6.slice/crio-b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63 WatchSource:0}: Error finding container b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63: Status 404 returned error can't find the container with id b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63 Oct 06 10:20:21 crc kubenswrapper[4989]: W1006 10:20:21.839605 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2fcb796_037b_403e_a59b_2d73a6e94de7.slice/crio-b8569d761c60fb239342700cd7817009ede126b4df69f18eade4ed29fde66e40 WatchSource:0}: Error finding container b8569d761c60fb239342700cd7817009ede126b4df69f18eade4ed29fde66e40: Status 404 returned error can't find the container with id b8569d761c60fb239342700cd7817009ede126b4df69f18eade4ed29fde66e40 Oct 06 10:20:21 crc kubenswrapper[4989]: I1006 10:20:21.852204 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-7j2m4"] Oct 06 10:20:21 crc kubenswrapper[4989]: I1006 10:20:21.856126 4989 scope.go:117] "RemoveContainer" containerID="0a4bb2572669056cfaa0c535b365a11c343e72325a6e6741efd32ebcdaa35195" Oct 06 10:20:21 crc kubenswrapper[4989]: I1006 10:20:21.885939 4989 scope.go:117] "RemoveContainer" containerID="63a6b31b4b99335a8445802e0285b1a876886407b9fa0b649033521e97818c7e" Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.429308 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-7j2m4" event={"ID":"a2fcb796-037b-403e-a59b-2d73a6e94de7","Type":"ContainerStarted","Data":"b8569d761c60fb239342700cd7817009ede126b4df69f18eade4ed29fde66e40"} Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.431562 4989 generic.go:334] "Generic (PLEG): container finished" podID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerID="03853485f39b9e0b613725d24e755ca49c884ebd65435baa89a21f7fb105cc30" exitCode=0 Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.431644 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-jcn7j" event={"ID":"46711b9f-dd60-4901-9f45-a3301b4ea9c6","Type":"ContainerDied","Data":"03853485f39b9e0b613725d24e755ca49c884ebd65435baa89a21f7fb105cc30"} Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.431689 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-jcn7j" event={"ID":"46711b9f-dd60-4901-9f45-a3301b4ea9c6","Type":"ContainerStarted","Data":"b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63"} Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.433343 4989 generic.go:334] "Generic (PLEG): container finished" podID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerID="93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b" exitCode=0 Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.433413 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gfppm" event={"ID":"fe4a4862-c2c3-4750-ad47-94dee1b0130b","Type":"ContainerDied","Data":"93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b"} Oct 06 10:20:22 crc kubenswrapper[4989]: I1006 10:20:22.436839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-p2zxx" event={"ID":"b3e3fae2-2565-4d6b-8a43-30e48c1f734c","Type":"ContainerStarted","Data":"accbd723c00c807be74f5bd2a954ad539746a889d9c876319e86cdcfcd5fc533"} Oct 06 10:20:23 crc kubenswrapper[4989]: I1006 10:20:23.449549 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-jcn7j" event={"ID":"46711b9f-dd60-4901-9f45-a3301b4ea9c6","Type":"ContainerStarted","Data":"fa254e9869eb48e481ff7d0cf58ee46418e55d6480f5242b927bde858d1bd328"} Oct 06 10:20:23 crc kubenswrapper[4989]: I1006 10:20:23.476438 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-jcn7j" podStartSLOduration=5.47641962 podStartE2EDuration="5.47641962s" podCreationTimestamp="2025-10-06 10:20:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:23.469710047 +0000 UTC m=+6074.259735637" watchObservedRunningTime="2025-10-06 10:20:23.47641962 +0000 UTC m=+6074.266445190" Oct 06 10:20:23 crc kubenswrapper[4989]: I1006 10:20:23.939466 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:20:23 crc kubenswrapper[4989]: E1006 10:20:23.940153 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.486329 4989 generic.go:334] "Generic (PLEG): container finished" podID="b3e3fae2-2565-4d6b-8a43-30e48c1f734c" containerID="accbd723c00c807be74f5bd2a954ad539746a889d9c876319e86cdcfcd5fc533" exitCode=0 Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.486424 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-p2zxx" event={"ID":"b3e3fae2-2565-4d6b-8a43-30e48c1f734c","Type":"ContainerDied","Data":"accbd723c00c807be74f5bd2a954ad539746a889d9c876319e86cdcfcd5fc533"} Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.526674 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-hkh7r"] Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.528908 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.531646 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.531988 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.548886 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-hkh7r"] Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624230 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-combined-ca-bundle\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624355 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data-merged\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624410 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-amphora-certs\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624457 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624508 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e308652c-1db7-49f7-bc52-c0b1e7818138-hm-ports\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.624902 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-scripts\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727231 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-combined-ca-bundle\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727333 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data-merged\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727388 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-amphora-certs\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727439 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727493 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e308652c-1db7-49f7-bc52-c0b1e7818138-hm-ports\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727536 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-scripts\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.727911 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data-merged\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.728535 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e308652c-1db7-49f7-bc52-c0b1e7818138-hm-ports\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.730993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-combined-ca-bundle\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.732021 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-config-data\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.734536 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-scripts\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.736064 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e308652c-1db7-49f7-bc52-c0b1e7818138-amphora-certs\") pod \"octavia-worker-hkh7r\" (UID: \"e308652c-1db7-49f7-bc52-c0b1e7818138\") " pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:24 crc kubenswrapper[4989]: I1006 10:20:24.871250 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.256649 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-wsxs8" Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.497461 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gfppm" event={"ID":"fe4a4862-c2c3-4750-ad47-94dee1b0130b","Type":"ContainerStarted","Data":"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5"} Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.499788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-p2zxx" event={"ID":"b3e3fae2-2565-4d6b-8a43-30e48c1f734c","Type":"ContainerStarted","Data":"7406fe8cd2cef6ae74e62f1aa0a0bfbbd683e3f8bc4849ddd53b75d5f870df54"} Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.500043 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.501950 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-7j2m4" event={"ID":"a2fcb796-037b-403e-a59b-2d73a6e94de7","Type":"ContainerStarted","Data":"2c2bba7a160252872962c938e2f0628f2f9c2e0b44dc54b92deb526bbf3276f3"} Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.519024 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-gfppm" podStartSLOduration=3.046802697 podStartE2EDuration="15.519007291s" podCreationTimestamp="2025-10-06 10:20:10 +0000 UTC" firstStartedPulling="2025-10-06 10:20:11.717116217 +0000 UTC m=+6062.507141797" lastFinishedPulling="2025-10-06 10:20:24.189320781 +0000 UTC m=+6074.979346391" observedRunningTime="2025-10-06 10:20:25.516859299 +0000 UTC m=+6076.306884899" watchObservedRunningTime="2025-10-06 10:20:25.519007291 +0000 UTC m=+6076.309032871" Oct 06 10:20:25 crc kubenswrapper[4989]: I1006 10:20:25.569117 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-p2zxx" podStartSLOduration=8.569098405 podStartE2EDuration="8.569098405s" podCreationTimestamp="2025-10-06 10:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:20:25.566691035 +0000 UTC m=+6076.356716605" watchObservedRunningTime="2025-10-06 10:20:25.569098405 +0000 UTC m=+6076.359123995" Oct 06 10:20:26 crc kubenswrapper[4989]: I1006 10:20:26.106617 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-hkh7r"] Oct 06 10:20:26 crc kubenswrapper[4989]: W1006 10:20:26.128755 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode308652c_1db7_49f7_bc52_c0b1e7818138.slice/crio-a84b79e817ee7be7a9b9eb52a8d44e3d542700d386f1ecc586a5798eb672ae5b WatchSource:0}: Error finding container a84b79e817ee7be7a9b9eb52a8d44e3d542700d386f1ecc586a5798eb672ae5b: Status 404 returned error can't find the container with id a84b79e817ee7be7a9b9eb52a8d44e3d542700d386f1ecc586a5798eb672ae5b Oct 06 10:20:26 crc kubenswrapper[4989]: I1006 10:20:26.511979 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-hkh7r" event={"ID":"e308652c-1db7-49f7-bc52-c0b1e7818138","Type":"ContainerStarted","Data":"a84b79e817ee7be7a9b9eb52a8d44e3d542700d386f1ecc586a5798eb672ae5b"} Oct 06 10:20:27 crc kubenswrapper[4989]: I1006 10:20:27.526076 4989 generic.go:334] "Generic (PLEG): container finished" podID="a2fcb796-037b-403e-a59b-2d73a6e94de7" containerID="2c2bba7a160252872962c938e2f0628f2f9c2e0b44dc54b92deb526bbf3276f3" exitCode=0 Oct 06 10:20:27 crc kubenswrapper[4989]: I1006 10:20:27.526181 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-7j2m4" event={"ID":"a2fcb796-037b-403e-a59b-2d73a6e94de7","Type":"ContainerDied","Data":"2c2bba7a160252872962c938e2f0628f2f9c2e0b44dc54b92deb526bbf3276f3"} Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.543162 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-hkh7r" event={"ID":"e308652c-1db7-49f7-bc52-c0b1e7818138","Type":"ContainerStarted","Data":"740a558668af9bb61e0231a913fee3bcc82130b34edf1b0ea990bc1a2104341a"} Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.545598 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-7j2m4" event={"ID":"a2fcb796-037b-403e-a59b-2d73a6e94de7","Type":"ContainerStarted","Data":"c370f173ae783cb11f335248865dc6c0a0e7c70686a1cea6f3ca37302d6b83da"} Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.545890 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.550702 4989 generic.go:334] "Generic (PLEG): container finished" podID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerID="fa254e9869eb48e481ff7d0cf58ee46418e55d6480f5242b927bde858d1bd328" exitCode=0 Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.550764 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-jcn7j" event={"ID":"46711b9f-dd60-4901-9f45-a3301b4ea9c6","Type":"ContainerDied","Data":"fa254e9869eb48e481ff7d0cf58ee46418e55d6480f5242b927bde858d1bd328"} Oct 06 10:20:28 crc kubenswrapper[4989]: I1006 10:20:28.614016 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-7j2m4" podStartSLOduration=6.265619643 podStartE2EDuration="8.61399848s" podCreationTimestamp="2025-10-06 10:20:20 +0000 UTC" firstStartedPulling="2025-10-06 10:20:21.842364175 +0000 UTC m=+6072.632389755" lastFinishedPulling="2025-10-06 10:20:24.190742972 +0000 UTC m=+6074.980768592" observedRunningTime="2025-10-06 10:20:28.605101024 +0000 UTC m=+6079.395126604" watchObservedRunningTime="2025-10-06 10:20:28.61399848 +0000 UTC m=+6079.404024060" Oct 06 10:20:29 crc kubenswrapper[4989]: I1006 10:20:29.567276 4989 generic.go:334] "Generic (PLEG): container finished" podID="e308652c-1db7-49f7-bc52-c0b1e7818138" containerID="740a558668af9bb61e0231a913fee3bcc82130b34edf1b0ea990bc1a2104341a" exitCode=0 Oct 06 10:20:29 crc kubenswrapper[4989]: I1006 10:20:29.567642 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-hkh7r" event={"ID":"e308652c-1db7-49f7-bc52-c0b1e7818138","Type":"ContainerDied","Data":"740a558668af9bb61e0231a913fee3bcc82130b34edf1b0ea990bc1a2104341a"} Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.064518 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.176369 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged\") pod \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.176420 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle\") pod \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.176483 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts\") pod \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.176607 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data\") pod \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\" (UID: \"46711b9f-dd60-4901-9f45-a3301b4ea9c6\") " Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.181888 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts" (OuterVolumeSpecName: "scripts") pod "46711b9f-dd60-4901-9f45-a3301b4ea9c6" (UID: "46711b9f-dd60-4901-9f45-a3301b4ea9c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.182577 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data" (OuterVolumeSpecName: "config-data") pod "46711b9f-dd60-4901-9f45-a3301b4ea9c6" (UID: "46711b9f-dd60-4901-9f45-a3301b4ea9c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.212461 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "46711b9f-dd60-4901-9f45-a3301b4ea9c6" (UID: "46711b9f-dd60-4901-9f45-a3301b4ea9c6"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.221120 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46711b9f-dd60-4901-9f45-a3301b4ea9c6" (UID: "46711b9f-dd60-4901-9f45-a3301b4ea9c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.283479 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.284000 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.284019 4989 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/46711b9f-dd60-4901-9f45-a3301b4ea9c6-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.284038 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46711b9f-dd60-4901-9f45-a3301b4ea9c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.583228 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-jcn7j" event={"ID":"46711b9f-dd60-4901-9f45-a3301b4ea9c6","Type":"ContainerDied","Data":"b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63"} Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.583275 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4e4983185888d0e35edba48c5e8532ead20fe0a870e977b101161ecc140dd63" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.583336 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-jcn7j" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.590356 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-hkh7r" event={"ID":"e308652c-1db7-49f7-bc52-c0b1e7818138","Type":"ContainerStarted","Data":"bab6e1a4f17806ee8d2e7922c8924fe90ac4adfdec2dd169eab63f8d7a896b8c"} Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.591550 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:30 crc kubenswrapper[4989]: I1006 10:20:30.629156 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-hkh7r" podStartSLOduration=5.262103972 podStartE2EDuration="6.62913521s" podCreationTimestamp="2025-10-06 10:20:24 +0000 UTC" firstStartedPulling="2025-10-06 10:20:26.136321576 +0000 UTC m=+6076.926347156" lastFinishedPulling="2025-10-06 10:20:27.503352784 +0000 UTC m=+6078.293378394" observedRunningTime="2025-10-06 10:20:30.615092235 +0000 UTC m=+6081.405117815" watchObservedRunningTime="2025-10-06 10:20:30.62913521 +0000 UTC m=+6081.419160790" Oct 06 10:20:32 crc kubenswrapper[4989]: I1006 10:20:32.565118 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-p2zxx" Oct 06 10:20:35 crc kubenswrapper[4989]: I1006 10:20:35.747669 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-7j2m4" Oct 06 10:20:36 crc kubenswrapper[4989]: I1006 10:20:36.936726 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:20:36 crc kubenswrapper[4989]: E1006 10:20:36.937149 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:20:39 crc kubenswrapper[4989]: I1006 10:20:39.909402 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-hkh7r" Oct 06 10:20:40 crc kubenswrapper[4989]: I1006 10:20:40.038811 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-drvvq"] Oct 06 10:20:40 crc kubenswrapper[4989]: I1006 10:20:40.049110 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-drvvq"] Oct 06 10:20:41 crc kubenswrapper[4989]: I1006 10:20:41.948811 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a2cde9-142d-4338-86f4-972c36750be2" path="/var/lib/kubelet/pods/f7a2cde9-142d-4338-86f4-972c36750be2/volumes" Oct 06 10:20:47 crc kubenswrapper[4989]: I1006 10:20:47.952485 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:20:47 crc kubenswrapper[4989]: E1006 10:20:47.953583 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:20:50 crc kubenswrapper[4989]: I1006 10:20:50.027478 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-9766-account-create-5lr7p"] Oct 06 10:20:50 crc kubenswrapper[4989]: I1006 10:20:50.041287 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-9766-account-create-5lr7p"] Oct 06 10:20:51 crc kubenswrapper[4989]: I1006 10:20:51.954131 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f17f9bc-804b-4244-bb7d-b297e1069061" path="/var/lib/kubelet/pods/3f17f9bc-804b-4244-bb7d-b297e1069061/volumes" Oct 06 10:20:57 crc kubenswrapper[4989]: I1006 10:20:57.053924 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-jhgpb"] Oct 06 10:20:57 crc kubenswrapper[4989]: I1006 10:20:57.065900 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-jhgpb"] Oct 06 10:20:57 crc kubenswrapper[4989]: I1006 10:20:57.948609 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d8844c1-5796-4d53-a114-856435a3d743" path="/var/lib/kubelet/pods/6d8844c1-5796-4d53-a114-856435a3d743/volumes" Oct 06 10:20:57 crc kubenswrapper[4989]: I1006 10:20:57.954043 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:57 crc kubenswrapper[4989]: I1006 10:20:57.954264 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-gfppm" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="octavia-amphora-httpd" containerID="cri-o://8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5" gracePeriod=30 Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.441175 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.610866 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image\") pod \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.611345 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config\") pod \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\" (UID: \"fe4a4862-c2c3-4750-ad47-94dee1b0130b\") " Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.680391 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fe4a4862-c2c3-4750-ad47-94dee1b0130b" (UID: "fe4a4862-c2c3-4750-ad47-94dee1b0130b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.704144 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "fe4a4862-c2c3-4750-ad47-94dee1b0130b" (UID: "fe4a4862-c2c3-4750-ad47-94dee1b0130b"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.714793 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe4a4862-c2c3-4750-ad47-94dee1b0130b-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.714828 4989 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/fe4a4862-c2c3-4750-ad47-94dee1b0130b-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.944159 4989 generic.go:334] "Generic (PLEG): container finished" podID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerID="8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5" exitCode=0 Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.944211 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gfppm" event={"ID":"fe4a4862-c2c3-4750-ad47-94dee1b0130b","Type":"ContainerDied","Data":"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5"} Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.944242 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-gfppm" event={"ID":"fe4a4862-c2c3-4750-ad47-94dee1b0130b","Type":"ContainerDied","Data":"204b1a5ae17c51afa395e88e1cbe94b92cd1473f7aec4a7a77b8114bf5b371f6"} Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.944262 4989 scope.go:117] "RemoveContainer" containerID="8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.944392 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-gfppm" Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.991974 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:58 crc kubenswrapper[4989]: I1006 10:20:58.995709 4989 scope.go:117] "RemoveContainer" containerID="93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b" Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.001790 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-gfppm"] Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.023923 4989 scope.go:117] "RemoveContainer" containerID="8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5" Oct 06 10:20:59 crc kubenswrapper[4989]: E1006 10:20:59.024966 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5\": container with ID starting with 8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5 not found: ID does not exist" containerID="8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5" Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.025022 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5"} err="failed to get container status \"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5\": rpc error: code = NotFound desc = could not find container \"8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5\": container with ID starting with 8f129f31d7c94b4e44e29d8db921b463f903fafce83b2304cbb0ac59e980e2b5 not found: ID does not exist" Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.025049 4989 scope.go:117] "RemoveContainer" containerID="93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b" Oct 06 10:20:59 crc kubenswrapper[4989]: E1006 10:20:59.025385 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b\": container with ID starting with 93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b not found: ID does not exist" containerID="93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b" Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.025408 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b"} err="failed to get container status \"93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b\": rpc error: code = NotFound desc = could not find container \"93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b\": container with ID starting with 93eb6f4056457001036f844cee3b4ea0ab0cc42f3cb70b740e57f863a1eecb3b not found: ID does not exist" Oct 06 10:20:59 crc kubenswrapper[4989]: I1006 10:20:59.957794 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" path="/var/lib/kubelet/pods/fe4a4862-c2c3-4750-ad47-94dee1b0130b/volumes" Oct 06 10:21:01 crc kubenswrapper[4989]: I1006 10:21:01.941894 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:21:01 crc kubenswrapper[4989]: E1006 10:21:01.942748 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.184078 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-hfddc"] Oct 06 10:21:03 crc kubenswrapper[4989]: E1006 10:21:03.184838 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerName="octavia-db-sync" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.184852 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerName="octavia-db-sync" Oct 06 10:21:03 crc kubenswrapper[4989]: E1006 10:21:03.184875 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerName="init" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.184881 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerName="init" Oct 06 10:21:03 crc kubenswrapper[4989]: E1006 10:21:03.184899 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="init" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.184905 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="init" Oct 06 10:21:03 crc kubenswrapper[4989]: E1006 10:21:03.184916 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="octavia-amphora-httpd" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.184923 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="octavia-amphora-httpd" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.185098 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4a4862-c2c3-4750-ad47-94dee1b0130b" containerName="octavia-amphora-httpd" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.185116 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" containerName="octavia-db-sync" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.187167 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.189935 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.193383 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-hfddc"] Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.316691 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/70a61785-944c-4912-8373-2e5161470a83-amphora-image\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.317381 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70a61785-944c-4912-8373-2e5161470a83-httpd-config\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.419024 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70a61785-944c-4912-8373-2e5161470a83-httpd-config\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.419367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/70a61785-944c-4912-8373-2e5161470a83-amphora-image\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.419791 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/70a61785-944c-4912-8373-2e5161470a83-amphora-image\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.424988 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/70a61785-944c-4912-8373-2e5161470a83-httpd-config\") pod \"octavia-image-upload-59f8cff499-hfddc\" (UID: \"70a61785-944c-4912-8373-2e5161470a83\") " pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:03 crc kubenswrapper[4989]: I1006 10:21:03.535924 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-hfddc" Oct 06 10:21:04 crc kubenswrapper[4989]: I1006 10:21:04.014874 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-hfddc"] Oct 06 10:21:05 crc kubenswrapper[4989]: I1006 10:21:05.035287 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-hfddc" event={"ID":"70a61785-944c-4912-8373-2e5161470a83","Type":"ContainerStarted","Data":"d128c200401bbc66dd8b84e959aceeccd0eab0f106b6e5df5ce2a515796eea0c"} Oct 06 10:21:05 crc kubenswrapper[4989]: I1006 10:21:05.035611 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-hfddc" event={"ID":"70a61785-944c-4912-8373-2e5161470a83","Type":"ContainerStarted","Data":"4e0c5b4dc10ed32adfbdd7fcad4794606f7a2163e10a2bf5701d6bb8131eeb00"} Oct 06 10:21:06 crc kubenswrapper[4989]: I1006 10:21:06.047294 4989 generic.go:334] "Generic (PLEG): container finished" podID="70a61785-944c-4912-8373-2e5161470a83" containerID="d128c200401bbc66dd8b84e959aceeccd0eab0f106b6e5df5ce2a515796eea0c" exitCode=0 Oct 06 10:21:06 crc kubenswrapper[4989]: I1006 10:21:06.047345 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-hfddc" event={"ID":"70a61785-944c-4912-8373-2e5161470a83","Type":"ContainerDied","Data":"d128c200401bbc66dd8b84e959aceeccd0eab0f106b6e5df5ce2a515796eea0c"} Oct 06 10:21:08 crc kubenswrapper[4989]: I1006 10:21:08.068357 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-hfddc" event={"ID":"70a61785-944c-4912-8373-2e5161470a83","Type":"ContainerStarted","Data":"14318db282cc50f40aaf61174a215c8bdb7e8feb161d8ed46afc523a41015b21"} Oct 06 10:21:08 crc kubenswrapper[4989]: I1006 10:21:08.086008 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-hfddc" podStartSLOduration=1.780351647 podStartE2EDuration="5.085983427s" podCreationTimestamp="2025-10-06 10:21:03 +0000 UTC" firstStartedPulling="2025-10-06 10:21:04.028539125 +0000 UTC m=+6114.818564705" lastFinishedPulling="2025-10-06 10:21:07.334170895 +0000 UTC m=+6118.124196485" observedRunningTime="2025-10-06 10:21:08.082780845 +0000 UTC m=+6118.872806435" watchObservedRunningTime="2025-10-06 10:21:08.085983427 +0000 UTC m=+6118.876009007" Oct 06 10:21:12 crc kubenswrapper[4989]: I1006 10:21:12.935788 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:21:12 crc kubenswrapper[4989]: E1006 10:21:12.936477 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:21:21 crc kubenswrapper[4989]: I1006 10:21:21.935316 4989 scope.go:117] "RemoveContainer" containerID="8afb79fc1865988b10e85b358785958e108891022b2b1c9dfab3860ef75bbf8e" Oct 06 10:21:21 crc kubenswrapper[4989]: I1006 10:21:21.980599 4989 scope.go:117] "RemoveContainer" containerID="4b7850684d23fdd264dd1b306607c85ba0abcb31b67d45fec01245852f5b575b" Oct 06 10:21:22 crc kubenswrapper[4989]: I1006 10:21:22.033202 4989 scope.go:117] "RemoveContainer" containerID="9d75d55d504f174a69d3c97fb792d6cad40c23a296dc43d69448a5835f8d3c23" Oct 06 10:21:22 crc kubenswrapper[4989]: I1006 10:21:22.068599 4989 scope.go:117] "RemoveContainer" containerID="cf768e88a3abfe99608b401828618af7036ef4b20ba2055e0b8522e189489eb9" Oct 06 10:21:22 crc kubenswrapper[4989]: I1006 10:21:22.088410 4989 scope.go:117] "RemoveContainer" containerID="cb45ef9cf20362313e0a31a718e8e7f6491936616a109088467309489e91655d" Oct 06 10:21:22 crc kubenswrapper[4989]: I1006 10:21:22.169512 4989 scope.go:117] "RemoveContainer" containerID="a3332bed5eb24cbc9d50b7e71dc1715e26fc0e8d875e96f3553f84a4cac7ee04" Oct 06 10:21:26 crc kubenswrapper[4989]: I1006 10:21:26.936593 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:21:26 crc kubenswrapper[4989]: E1006 10:21:26.937275 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:21:28 crc kubenswrapper[4989]: I1006 10:21:28.035297 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-5j6jz"] Oct 06 10:21:28 crc kubenswrapper[4989]: I1006 10:21:28.048877 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-5j6jz"] Oct 06 10:21:29 crc kubenswrapper[4989]: I1006 10:21:29.947921 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87efa56a-fb99-42e8-91d6-680def8278ed" path="/var/lib/kubelet/pods/87efa56a-fb99-42e8-91d6-680def8278ed/volumes" Oct 06 10:21:38 crc kubenswrapper[4989]: I1006 10:21:38.053554 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86e0-account-create-rc22c"] Oct 06 10:21:38 crc kubenswrapper[4989]: I1006 10:21:38.065909 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-86e0-account-create-rc22c"] Oct 06 10:21:39 crc kubenswrapper[4989]: I1006 10:21:39.941083 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:21:39 crc kubenswrapper[4989]: E1006 10:21:39.941548 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:21:39 crc kubenswrapper[4989]: I1006 10:21:39.952055 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5646c4e1-673f-4571-9d34-9558a1d7e0ce" path="/var/lib/kubelet/pods/5646c4e1-673f-4571-9d34-9558a1d7e0ce/volumes" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.064583 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.073536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.079742 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-4nqng" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.080211 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.080479 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.082613 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.090855 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.129249 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.129692 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-log" containerID="cri-o://60ae53ed77dea2379dd474804638c0824f7e05563aaabb85a9c29c98eb00fa88" gracePeriod=30 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.129741 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-httpd" containerID="cri-o://77f84f997ad1f7e98a8c6a63330f4bbb258d682434bff7e7e48b82b66af03cf3" gracePeriod=30 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.163931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6clf\" (UniqueName: \"kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.163980 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.164118 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.164156 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.164237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.228698 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.228999 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-log" containerID="cri-o://3aeba07768ba71f5efff33027feece42bf9054252e82470640a86584b7aa923f" gracePeriod=30 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.229172 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-httpd" containerID="cri-o://eb83642e14061395a0ab65f70817c6ab2e76e892d431600bbb05ad2460bd2101" gracePeriod=30 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.250768 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.252467 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.268785 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.268835 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.268900 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.268927 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6clf\" (UniqueName: \"kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.268943 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.269419 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.269957 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.271335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.293704 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.297498 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.320202 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6clf\" (UniqueName: \"kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf\") pod \"horizon-b9f98794f-t426s\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.370748 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.370800 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksb2k\" (UniqueName: \"kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.370824 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.370883 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.370911 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.399132 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.473550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.473607 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksb2k\" (UniqueName: \"kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.473633 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.473718 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.473746 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.474857 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.475311 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.476377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.478473 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.507399 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksb2k\" (UniqueName: \"kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k\") pod \"horizon-6cc44c5b49-tsp7j\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.532504 4989 generic.go:334] "Generic (PLEG): container finished" podID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerID="60ae53ed77dea2379dd474804638c0824f7e05563aaabb85a9c29c98eb00fa88" exitCode=143 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.532563 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerDied","Data":"60ae53ed77dea2379dd474804638c0824f7e05563aaabb85a9c29c98eb00fa88"} Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.547486 4989 generic.go:334] "Generic (PLEG): container finished" podID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerID="3aeba07768ba71f5efff33027feece42bf9054252e82470640a86584b7aa923f" exitCode=143 Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.547526 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerDied","Data":"3aeba07768ba71f5efff33027feece42bf9054252e82470640a86584b7aa923f"} Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.712047 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.872721 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.922248 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.944078 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:46 crc kubenswrapper[4989]: I1006 10:21:46.948610 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:21:47 crc kubenswrapper[4989]: W1006 10:21:47.002725 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48abde4f_40c9_42e2_aaf7_a26899e03a41.slice/crio-c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e WatchSource:0}: Error finding container c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e: Status 404 returned error can't find the container with id c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.008135 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.045971 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.056303 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lhdlc"] Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.061605 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lhdlc"] Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.089330 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.089370 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.089531 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.089593 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vdz\" (UniqueName: \"kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.089617 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.192097 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.192184 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vdz\" (UniqueName: \"kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.193057 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.193195 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.193222 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.193470 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.194076 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.194791 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.201106 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.218150 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vdz\" (UniqueName: \"kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz\") pod \"horizon-6c6ddd9d9c-kz8rv\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:47 crc kubenswrapper[4989]: I1006 10:21:47.287091 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:48 crc kubenswrapper[4989]: I1006 10:21:47.561283 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerStarted","Data":"c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e"} Oct 06 10:21:48 crc kubenswrapper[4989]: I1006 10:21:47.562900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerStarted","Data":"9789ccb3d66d9df4b14cbaf5893118b88566c4ddede33bd0c2d12c212f5cdc2e"} Oct 06 10:21:48 crc kubenswrapper[4989]: I1006 10:21:47.956848 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="678f61cd-297a-469e-b897-ec94ec29daf0" path="/var/lib/kubelet/pods/678f61cd-297a-469e-b897-ec94ec29daf0/volumes" Oct 06 10:21:48 crc kubenswrapper[4989]: W1006 10:21:48.573719 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd280e305_5e2d_4e2f_a421_1e0c93c598ab.slice/crio-8f191774d71d92515dbd87ae782f83d7412ecd233fe34bf13de2677de1f045a5 WatchSource:0}: Error finding container 8f191774d71d92515dbd87ae782f83d7412ecd233fe34bf13de2677de1f045a5: Status 404 returned error can't find the container with id 8f191774d71d92515dbd87ae782f83d7412ecd233fe34bf13de2677de1f045a5 Oct 06 10:21:48 crc kubenswrapper[4989]: I1006 10:21:48.575832 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:21:49 crc kubenswrapper[4989]: I1006 10:21:49.584329 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerStarted","Data":"8f191774d71d92515dbd87ae782f83d7412ecd233fe34bf13de2677de1f045a5"} Oct 06 10:21:50 crc kubenswrapper[4989]: I1006 10:21:50.598673 4989 generic.go:334] "Generic (PLEG): container finished" podID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerID="eb83642e14061395a0ab65f70817c6ab2e76e892d431600bbb05ad2460bd2101" exitCode=0 Oct 06 10:21:50 crc kubenswrapper[4989]: I1006 10:21:50.599083 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerDied","Data":"eb83642e14061395a0ab65f70817c6ab2e76e892d431600bbb05ad2460bd2101"} Oct 06 10:21:51 crc kubenswrapper[4989]: I1006 10:21:51.609800 4989 generic.go:334] "Generic (PLEG): container finished" podID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerID="77f84f997ad1f7e98a8c6a63330f4bbb258d682434bff7e7e48b82b66af03cf3" exitCode=0 Oct 06 10:21:51 crc kubenswrapper[4989]: I1006 10:21:51.609854 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerDied","Data":"77f84f997ad1f7e98a8c6a63330f4bbb258d682434bff7e7e48b82b66af03cf3"} Oct 06 10:21:52 crc kubenswrapper[4989]: I1006 10:21:52.936004 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:21:52 crc kubenswrapper[4989]: E1006 10:21:52.936672 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:21:53 crc kubenswrapper[4989]: I1006 10:21:53.967140 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.41:9292/healthcheck\": dial tcp 10.217.1.41:9292: connect: connection refused" Oct 06 10:21:53 crc kubenswrapper[4989]: I1006 10:21:53.967435 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.41:9292/healthcheck\": dial tcp 10.217.1.41:9292: connect: connection refused" Oct 06 10:21:54 crc kubenswrapper[4989]: I1006 10:21:54.972320 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085049 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085091 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv9vs\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085142 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085170 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085273 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085293 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.085358 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts\") pod \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\" (UID: \"34b488a4-65f3-446c-9e0b-7c78fb8ef126\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.086933 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.087248 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs" (OuterVolumeSpecName: "logs") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.091582 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs" (OuterVolumeSpecName: "kube-api-access-nv9vs") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "kube-api-access-nv9vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.093206 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts" (OuterVolumeSpecName: "scripts") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.097391 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph" (OuterVolumeSpecName: "ceph") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.133176 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.156231 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data" (OuterVolumeSpecName: "config-data") pod "34b488a4-65f3-446c-9e0b-7c78fb8ef126" (UID: "34b488a4-65f3-446c-9e0b-7c78fb8ef126"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.187369 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.187885 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv9vs\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-kube-api-access-nv9vs\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.188093 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.188214 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.188316 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34b488a4-65f3-446c-9e0b-7c78fb8ef126-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.188412 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/34b488a4-65f3-446c-9e0b-7c78fb8ef126-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.188501 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b488a4-65f3-446c-9e0b-7c78fb8ef126-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.660055 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerStarted","Data":"1bc81d3a014ed55b5dfe4a0a40e3c7b54446aa020e598da857f402e485de772b"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.660100 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerStarted","Data":"7ca8944d8a3552d46c08819280437e46fb83766e451cda7aa935c60704cdf59c"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.660112 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cc44c5b49-tsp7j" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon-log" containerID="cri-o://7ca8944d8a3552d46c08819280437e46fb83766e451cda7aa935c60704cdf59c" gracePeriod=30 Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.660210 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cc44c5b49-tsp7j" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon" containerID="cri-o://1bc81d3a014ed55b5dfe4a0a40e3c7b54446aa020e598da857f402e485de772b" gracePeriod=30 Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.681551 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerStarted","Data":"920efe9d85bc10aecfa0897062089e0e9b85f7bdbb880e2d66ddb48aec0937ba"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.681597 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerStarted","Data":"f8c909ea9a4843ca480af0ad547dcb57df96bd5ebf3752bb7606f73681afcb86"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.686320 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerStarted","Data":"5b6637f416ca8b50111683091d6ad8612c48384d8c4626c6d2b2fd07d6131d51"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.686447 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerStarted","Data":"e0fbd6d1dfe04b8a18573fe65060f2861d4c2d923909672edeabe24974e944a5"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.688749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34b488a4-65f3-446c-9e0b-7c78fb8ef126","Type":"ContainerDied","Data":"4d82424efeeddf6e34d471014208d6f1b2d69d3c5088399d880d08f1c8855e82"} Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.688880 4989 scope.go:117] "RemoveContainer" containerID="77f84f997ad1f7e98a8c6a63330f4bbb258d682434bff7e7e48b82b66af03cf3" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.689080 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.692777 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cc44c5b49-tsp7j" podStartSLOduration=2.2188903 podStartE2EDuration="9.692760537s" podCreationTimestamp="2025-10-06 10:21:46 +0000 UTC" firstStartedPulling="2025-10-06 10:21:47.051011864 +0000 UTC m=+6157.841037444" lastFinishedPulling="2025-10-06 10:21:54.524882081 +0000 UTC m=+6165.314907681" observedRunningTime="2025-10-06 10:21:55.681858412 +0000 UTC m=+6166.471884012" watchObservedRunningTime="2025-10-06 10:21:55.692760537 +0000 UTC m=+6166.482786117" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.707558 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-b9f98794f-t426s" podStartSLOduration=2.197493043 podStartE2EDuration="9.707537513s" podCreationTimestamp="2025-10-06 10:21:46 +0000 UTC" firstStartedPulling="2025-10-06 10:21:47.012072331 +0000 UTC m=+6157.802097911" lastFinishedPulling="2025-10-06 10:21:54.522116801 +0000 UTC m=+6165.312142381" observedRunningTime="2025-10-06 10:21:55.702506328 +0000 UTC m=+6166.492531918" watchObservedRunningTime="2025-10-06 10:21:55.707537513 +0000 UTC m=+6166.497563083" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.770092 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podStartSLOduration=3.777051896 podStartE2EDuration="9.770073795s" podCreationTimestamp="2025-10-06 10:21:46 +0000 UTC" firstStartedPulling="2025-10-06 10:21:48.575815228 +0000 UTC m=+6159.365840808" lastFinishedPulling="2025-10-06 10:21:54.568837127 +0000 UTC m=+6165.358862707" observedRunningTime="2025-10-06 10:21:55.722050731 +0000 UTC m=+6166.512076311" watchObservedRunningTime="2025-10-06 10:21:55.770073795 +0000 UTC m=+6166.560099375" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.770814 4989 scope.go:117] "RemoveContainer" containerID="60ae53ed77dea2379dd474804638c0824f7e05563aaabb85a9c29c98eb00fa88" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.783259 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.807683 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.821611 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.831288 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:55 crc kubenswrapper[4989]: E1006 10:21:55.833927 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.833950 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: E1006 10:21:55.833981 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.833987 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: E1006 10:21:55.833998 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834005 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: E1006 10:21:55.834028 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834033 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834213 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834230 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-httpd" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834242 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.834262 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" containerName="glance-log" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.835321 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.837228 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.895701 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.901608 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.901797 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.901854 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.901882 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.901966 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmhd5\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902030 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902093 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle\") pod \"44aa1eb6-a9b6-422f-a91f-27ef20555668\" (UID: \"44aa1eb6-a9b6-422f-a91f-27ef20555668\") " Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902403 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-ceph\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902436 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902479 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902500 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm5j9\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-kube-api-access-tm5j9\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-logs\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902691 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.902727 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.909161 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs" (OuterVolumeSpecName: "logs") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.911767 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.915557 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph" (OuterVolumeSpecName: "ceph") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.920927 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5" (OuterVolumeSpecName: "kube-api-access-hmhd5") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "kube-api-access-hmhd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.945062 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts" (OuterVolumeSpecName: "scripts") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.965767 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:55 crc kubenswrapper[4989]: I1006 10:21:55.985391 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b488a4-65f3-446c-9e0b-7c78fb8ef126" path="/var/lib/kubelet/pods/34b488a4-65f3-446c-9e0b-7c78fb8ef126/volumes" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004719 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004769 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004804 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-ceph\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004831 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004877 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.004893 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm5j9\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-kube-api-access-tm5j9\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005025 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-logs\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005090 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005100 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005108 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005117 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmhd5\" (UniqueName: \"kubernetes.io/projected/44aa1eb6-a9b6-422f-a91f-27ef20555668-kube-api-access-hmhd5\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005127 4989 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/44aa1eb6-a9b6-422f-a91f-27ef20555668-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.005135 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.012920 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-logs\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.016322 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-config-data\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.020374 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.020558 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed83ef55-63eb-4626-9661-d64e098a04db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.020936 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed83ef55-63eb-4626-9661-d64e098a04db-scripts\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.024102 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-ceph\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.037005 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm5j9\" (UniqueName: \"kubernetes.io/projected/ed83ef55-63eb-4626-9661-d64e098a04db-kube-api-access-tm5j9\") pod \"glance-default-external-api-0\" (UID: \"ed83ef55-63eb-4626-9661-d64e098a04db\") " pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.045043 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data" (OuterVolumeSpecName: "config-data") pod "44aa1eb6-a9b6-422f-a91f-27ef20555668" (UID: "44aa1eb6-a9b6-422f-a91f-27ef20555668"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.106893 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44aa1eb6-a9b6-422f-a91f-27ef20555668-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.193276 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.408499 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.408800 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.588062 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.703648 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"44aa1eb6-a9b6-422f-a91f-27ef20555668","Type":"ContainerDied","Data":"ecd4ff97f82c465d72ac8dd147790a03e2a8b9dc8f6aa7e13e78ce1692045f8a"} Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.703706 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.703720 4989 scope.go:117] "RemoveContainer" containerID="eb83642e14061395a0ab65f70817c6ab2e76e892d431600bbb05ad2460bd2101" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.712580 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed83ef55-63eb-4626-9661-d64e098a04db","Type":"ContainerStarted","Data":"f0aa653fd541a263da628629f234f69e0ca3346ad619f7b3f00352b874e2c2ce"} Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.712807 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.744852 4989 scope.go:117] "RemoveContainer" containerID="3aeba07768ba71f5efff33027feece42bf9054252e82470640a86584b7aa923f" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.746569 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.762212 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.772019 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.774174 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.781208 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.802185 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.829913 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830008 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830045 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830103 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830148 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830191 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.830219 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5nzb\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-kube-api-access-m5nzb\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.932893 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.932965 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.932983 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.933011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.933045 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.933077 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.933095 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5nzb\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-kube-api-access-m5nzb\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.936161 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-logs\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.941880 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.945142 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.948882 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.949416 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.952487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:56 crc kubenswrapper[4989]: I1006 10:21:56.962400 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5nzb\" (UniqueName: \"kubernetes.io/projected/a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed-kube-api-access-m5nzb\") pod \"glance-default-internal-api-0\" (UID: \"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed\") " pod="openstack/glance-default-internal-api-0" Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.111837 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.288826 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.289292 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.703966 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 06 10:21:57 crc kubenswrapper[4989]: W1006 10:21:57.716031 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3c491b5_c32c_4e2e_bbf1_5ea7bd33e1ed.slice/crio-6a87d93c1cf7666e216a8a73277fb97d9f24918ca03539895b0b637d8ae1eb12 WatchSource:0}: Error finding container 6a87d93c1cf7666e216a8a73277fb97d9f24918ca03539895b0b637d8ae1eb12: Status 404 returned error can't find the container with id 6a87d93c1cf7666e216a8a73277fb97d9f24918ca03539895b0b637d8ae1eb12 Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.722533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed83ef55-63eb-4626-9661-d64e098a04db","Type":"ContainerStarted","Data":"a797f7abe1220456077984fe7e85f46b3b55a18082d9a950afc3eba645d566ad"} Oct 06 10:21:57 crc kubenswrapper[4989]: I1006 10:21:57.953102 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44aa1eb6-a9b6-422f-a91f-27ef20555668" path="/var/lib/kubelet/pods/44aa1eb6-a9b6-422f-a91f-27ef20555668/volumes" Oct 06 10:21:58 crc kubenswrapper[4989]: I1006 10:21:58.741287 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ed83ef55-63eb-4626-9661-d64e098a04db","Type":"ContainerStarted","Data":"07564b78e4822e775c38a87876ae9678d3b2dc9c5ccf02e19cb1cbad4448be45"} Oct 06 10:21:58 crc kubenswrapper[4989]: I1006 10:21:58.750069 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed","Type":"ContainerStarted","Data":"1833420169d67761743028d17d2317abddac6f22e6e87b638f7dfc62feeeed3b"} Oct 06 10:21:58 crc kubenswrapper[4989]: I1006 10:21:58.750122 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed","Type":"ContainerStarted","Data":"6a87d93c1cf7666e216a8a73277fb97d9f24918ca03539895b0b637d8ae1eb12"} Oct 06 10:21:58 crc kubenswrapper[4989]: I1006 10:21:58.766379 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.766344428 podStartE2EDuration="3.766344428s" podCreationTimestamp="2025-10-06 10:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:21:58.762518868 +0000 UTC m=+6169.552544448" watchObservedRunningTime="2025-10-06 10:21:58.766344428 +0000 UTC m=+6169.556370008" Oct 06 10:21:59 crc kubenswrapper[4989]: I1006 10:21:59.767263 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed","Type":"ContainerStarted","Data":"d454b56927b8de5a0b4c95454d72e4333dd91792ce1b11a9b71213af4f725c62"} Oct 06 10:21:59 crc kubenswrapper[4989]: I1006 10:21:59.802867 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.802844187 podStartE2EDuration="3.802844187s" podCreationTimestamp="2025-10-06 10:21:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:21:59.790284075 +0000 UTC m=+6170.580309715" watchObservedRunningTime="2025-10-06 10:21:59.802844187 +0000 UTC m=+6170.592869777" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.194134 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.194755 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.228460 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.251608 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.401621 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.847467 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:22:06 crc kubenswrapper[4989]: I1006 10:22:06.847521 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.112693 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.112743 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.152902 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.187737 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.292392 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.861440 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.861509 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:07 crc kubenswrapper[4989]: I1006 10:22:07.937267 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:22:07 crc kubenswrapper[4989]: E1006 10:22:07.938009 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:22:08 crc kubenswrapper[4989]: I1006 10:22:08.938323 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:22:08 crc kubenswrapper[4989]: I1006 10:22:08.938412 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 10:22:08 crc kubenswrapper[4989]: I1006 10:22:08.938754 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 06 10:22:09 crc kubenswrapper[4989]: I1006 10:22:09.918561 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:09 crc kubenswrapper[4989]: I1006 10:22:09.919010 4989 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 06 10:22:09 crc kubenswrapper[4989]: I1006 10:22:09.974270 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 06 10:22:18 crc kubenswrapper[4989]: I1006 10:22:18.116092 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:22:19 crc kubenswrapper[4989]: I1006 10:22:19.079630 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:22:19 crc kubenswrapper[4989]: I1006 10:22:19.760260 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:22:20 crc kubenswrapper[4989]: I1006 10:22:20.904592 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:22:20 crc kubenswrapper[4989]: I1006 10:22:20.993279 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:22:20 crc kubenswrapper[4989]: I1006 10:22:20.993522 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon-log" containerID="cri-o://f8c909ea9a4843ca480af0ad547dcb57df96bd5ebf3752bb7606f73681afcb86" gracePeriod=30 Oct 06 10:22:20 crc kubenswrapper[4989]: I1006 10:22:20.993808 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" containerID="cri-o://920efe9d85bc10aecfa0897062089e0e9b85f7bdbb880e2d66ddb48aec0937ba" gracePeriod=30 Oct 06 10:22:21 crc kubenswrapper[4989]: I1006 10:22:21.936342 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:22:21 crc kubenswrapper[4989]: E1006 10:22:21.936834 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:22:22 crc kubenswrapper[4989]: I1006 10:22:22.323939 4989 scope.go:117] "RemoveContainer" containerID="6fe6755f0bb0e9cc7ef1281e2d66bc095e6147a885eb3c3aba97bcb22037db23" Oct 06 10:22:22 crc kubenswrapper[4989]: I1006 10:22:22.351504 4989 scope.go:117] "RemoveContainer" containerID="423056df6e55c2701a50b591bc8a5733a660ec921c66153e04b23600d8f28358" Oct 06 10:22:22 crc kubenswrapper[4989]: I1006 10:22:22.406266 4989 scope.go:117] "RemoveContainer" containerID="7a2ff0e4f886535744336ed8e4c102b8504eff528a009a998f52025f453c7ddf" Oct 06 10:22:25 crc kubenswrapper[4989]: I1006 10:22:25.083714 4989 generic.go:334] "Generic (PLEG): container finished" podID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerID="920efe9d85bc10aecfa0897062089e0e9b85f7bdbb880e2d66ddb48aec0937ba" exitCode=0 Oct 06 10:22:25 crc kubenswrapper[4989]: I1006 10:22:25.083776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerDied","Data":"920efe9d85bc10aecfa0897062089e0e9b85f7bdbb880e2d66ddb48aec0937ba"} Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.095534 4989 generic.go:334] "Generic (PLEG): container finished" podID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerID="1bc81d3a014ed55b5dfe4a0a40e3c7b54446aa020e598da857f402e485de772b" exitCode=137 Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.095989 4989 generic.go:334] "Generic (PLEG): container finished" podID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerID="7ca8944d8a3552d46c08819280437e46fb83766e451cda7aa935c60704cdf59c" exitCode=137 Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.095627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerDied","Data":"1bc81d3a014ed55b5dfe4a0a40e3c7b54446aa020e598da857f402e485de772b"} Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.096070 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerDied","Data":"7ca8944d8a3552d46c08819280437e46fb83766e451cda7aa935c60704cdf59c"} Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.096084 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cc44c5b49-tsp7j" event={"ID":"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017","Type":"ContainerDied","Data":"9789ccb3d66d9df4b14cbaf5893118b88566c4ddede33bd0c2d12c212f5cdc2e"} Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.096096 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9789ccb3d66d9df4b14cbaf5893118b88566c4ddede33bd0c2d12c212f5cdc2e" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.121501 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.228137 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts\") pod \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.228188 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data\") pod \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.228296 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksb2k\" (UniqueName: \"kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k\") pod \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.228390 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs\") pod \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.228475 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key\") pod \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\" (UID: \"64c6c279-f2c6-4387-bcb6-b8ffd4a8c017\") " Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.229590 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs" (OuterVolumeSpecName: "logs") pod "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" (UID: "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.233911 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" (UID: "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.239914 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k" (OuterVolumeSpecName: "kube-api-access-ksb2k") pod "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" (UID: "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017"). InnerVolumeSpecName "kube-api-access-ksb2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.259240 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data" (OuterVolumeSpecName: "config-data") pod "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" (UID: "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.270121 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts" (OuterVolumeSpecName: "scripts") pod "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" (UID: "64c6c279-f2c6-4387-bcb6-b8ffd4a8c017"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.331260 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.331295 4989 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.331307 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.331317 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.331352 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksb2k\" (UniqueName: \"kubernetes.io/projected/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017-kube-api-access-ksb2k\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:26 crc kubenswrapper[4989]: I1006 10:22:26.399787 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 06 10:22:27 crc kubenswrapper[4989]: I1006 10:22:27.107848 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cc44c5b49-tsp7j" Oct 06 10:22:27 crc kubenswrapper[4989]: I1006 10:22:27.145807 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:22:27 crc kubenswrapper[4989]: I1006 10:22:27.154332 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6cc44c5b49-tsp7j"] Oct 06 10:22:27 crc kubenswrapper[4989]: I1006 10:22:27.953251 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" path="/var/lib/kubelet/pods/64c6c279-f2c6-4387-bcb6-b8ffd4a8c017/volumes" Oct 06 10:22:30 crc kubenswrapper[4989]: I1006 10:22:30.063397 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-d2dmk"] Oct 06 10:22:30 crc kubenswrapper[4989]: I1006 10:22:30.079452 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-d2dmk"] Oct 06 10:22:31 crc kubenswrapper[4989]: I1006 10:22:31.955430 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90ca868d-9f29-4ec0-8ca1-dbb6403bebb8" path="/var/lib/kubelet/pods/90ca868d-9f29-4ec0-8ca1-dbb6403bebb8/volumes" Oct 06 10:22:34 crc kubenswrapper[4989]: I1006 10:22:34.935846 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:22:34 crc kubenswrapper[4989]: E1006 10:22:34.936606 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:22:36 crc kubenswrapper[4989]: I1006 10:22:36.400461 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 06 10:22:40 crc kubenswrapper[4989]: I1006 10:22:40.082498 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fe7d-account-create-qsrmz"] Oct 06 10:22:40 crc kubenswrapper[4989]: I1006 10:22:40.094539 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fe7d-account-create-qsrmz"] Oct 06 10:22:41 crc kubenswrapper[4989]: I1006 10:22:41.951721 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21f0d9a4-084a-42e0-86ae-4dba3793472b" path="/var/lib/kubelet/pods/21f0d9a4-084a-42e0-86ae-4dba3793472b/volumes" Oct 06 10:22:45 crc kubenswrapper[4989]: I1006 10:22:45.936501 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:22:45 crc kubenswrapper[4989]: E1006 10:22:45.937289 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:22:46 crc kubenswrapper[4989]: I1006 10:22:46.400619 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-b9f98794f-t426s" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Oct 06 10:22:46 crc kubenswrapper[4989]: I1006 10:22:46.401130 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:22:49 crc kubenswrapper[4989]: I1006 10:22:49.034596 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-rbnm2"] Oct 06 10:22:49 crc kubenswrapper[4989]: I1006 10:22:49.047955 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-rbnm2"] Oct 06 10:22:49 crc kubenswrapper[4989]: I1006 10:22:49.954238 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ee627a-7a93-4ffc-b36f-f8c4b0a06839" path="/var/lib/kubelet/pods/32ee627a-7a93-4ffc-b36f-f8c4b0a06839/volumes" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.396512 4989 generic.go:334] "Generic (PLEG): container finished" podID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerID="f8c909ea9a4843ca480af0ad547dcb57df96bd5ebf3752bb7606f73681afcb86" exitCode=137 Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.396563 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerDied","Data":"f8c909ea9a4843ca480af0ad547dcb57df96bd5ebf3752bb7606f73681afcb86"} Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.397054 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b9f98794f-t426s" event={"ID":"48abde4f-40c9-42e2-aaf7-a26899e03a41","Type":"ContainerDied","Data":"c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e"} Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.397091 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8697eb37ffad86ea8464247bcc0f3da395344f646a9c93df33c7419a793932e" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.447702 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.539192 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs\") pod \"48abde4f-40c9-42e2-aaf7-a26899e03a41\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.539412 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6clf\" (UniqueName: \"kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf\") pod \"48abde4f-40c9-42e2-aaf7-a26899e03a41\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.539493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key\") pod \"48abde4f-40c9-42e2-aaf7-a26899e03a41\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.539552 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts\") pod \"48abde4f-40c9-42e2-aaf7-a26899e03a41\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.539592 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data\") pod \"48abde4f-40c9-42e2-aaf7-a26899e03a41\" (UID: \"48abde4f-40c9-42e2-aaf7-a26899e03a41\") " Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.540149 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs" (OuterVolumeSpecName: "logs") pod "48abde4f-40c9-42e2-aaf7-a26899e03a41" (UID: "48abde4f-40c9-42e2-aaf7-a26899e03a41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.544890 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "48abde4f-40c9-42e2-aaf7-a26899e03a41" (UID: "48abde4f-40c9-42e2-aaf7-a26899e03a41"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.546006 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf" (OuterVolumeSpecName: "kube-api-access-d6clf") pod "48abde4f-40c9-42e2-aaf7-a26899e03a41" (UID: "48abde4f-40c9-42e2-aaf7-a26899e03a41"). InnerVolumeSpecName "kube-api-access-d6clf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.562969 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data" (OuterVolumeSpecName: "config-data") pod "48abde4f-40c9-42e2-aaf7-a26899e03a41" (UID: "48abde4f-40c9-42e2-aaf7-a26899e03a41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.570010 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts" (OuterVolumeSpecName: "scripts") pod "48abde4f-40c9-42e2-aaf7-a26899e03a41" (UID: "48abde4f-40c9-42e2-aaf7-a26899e03a41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.642331 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48abde4f-40c9-42e2-aaf7-a26899e03a41-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.642573 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6clf\" (UniqueName: \"kubernetes.io/projected/48abde4f-40c9-42e2-aaf7-a26899e03a41-kube-api-access-d6clf\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.642725 4989 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/48abde4f-40c9-42e2-aaf7-a26899e03a41-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.642836 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:51 crc kubenswrapper[4989]: I1006 10:22:51.642936 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/48abde4f-40c9-42e2-aaf7-a26899e03a41-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:52 crc kubenswrapper[4989]: I1006 10:22:52.409873 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b9f98794f-t426s" Oct 06 10:22:52 crc kubenswrapper[4989]: I1006 10:22:52.447003 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:22:52 crc kubenswrapper[4989]: I1006 10:22:52.462570 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b9f98794f-t426s"] Oct 06 10:22:53 crc kubenswrapper[4989]: I1006 10:22:53.948991 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" path="/var/lib/kubelet/pods/48abde4f-40c9-42e2-aaf7-a26899e03a41/volumes" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.798542 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f868b9f5c-rq2jl"] Oct 06 10:22:54 crc kubenswrapper[4989]: E1006 10:22:54.799328 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799352 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: E1006 10:22:54.799369 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799378 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: E1006 10:22:54.799408 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799417 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: E1006 10:22:54.799453 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799461 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799732 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799762 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="48abde4f-40c9-42e2-aaf7-a26899e03a41" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799776 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon-log" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.799786 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c6c279-f2c6-4387-bcb6-b8ffd4a8c017" containerName="horizon" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.801025 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.815013 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f868b9f5c-rq2jl"] Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.916725 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-scripts\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.916792 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157d6322-0be6-423b-8276-c789369b07d2-logs\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.916870 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlg96\" (UniqueName: \"kubernetes.io/projected/157d6322-0be6-423b-8276-c789369b07d2-kube-api-access-dlg96\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.916942 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/157d6322-0be6-423b-8276-c789369b07d2-horizon-secret-key\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:54 crc kubenswrapper[4989]: I1006 10:22:54.916966 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-config-data\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.018134 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/157d6322-0be6-423b-8276-c789369b07d2-horizon-secret-key\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.018182 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-config-data\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.018207 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-scripts\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.018246 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157d6322-0be6-423b-8276-c789369b07d2-logs\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.018319 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlg96\" (UniqueName: \"kubernetes.io/projected/157d6322-0be6-423b-8276-c789369b07d2-kube-api-access-dlg96\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.019317 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/157d6322-0be6-423b-8276-c789369b07d2-logs\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.019382 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-scripts\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.019461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/157d6322-0be6-423b-8276-c789369b07d2-config-data\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.026763 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/157d6322-0be6-423b-8276-c789369b07d2-horizon-secret-key\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.037603 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlg96\" (UniqueName: \"kubernetes.io/projected/157d6322-0be6-423b-8276-c789369b07d2-kube-api-access-dlg96\") pod \"horizon-f868b9f5c-rq2jl\" (UID: \"157d6322-0be6-423b-8276-c789369b07d2\") " pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.120475 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:22:55 crc kubenswrapper[4989]: I1006 10:22:55.599720 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f868b9f5c-rq2jl"] Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.075647 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-2284z"] Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.077390 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2284z" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.085075 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2284z"] Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.240709 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76s5h\" (UniqueName: \"kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h\") pod \"heat-db-create-2284z\" (UID: \"1e70b45b-b8f8-43e9-aec7-fc621835edf1\") " pod="openstack/heat-db-create-2284z" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.342204 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76s5h\" (UniqueName: \"kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h\") pod \"heat-db-create-2284z\" (UID: \"1e70b45b-b8f8-43e9-aec7-fc621835edf1\") " pod="openstack/heat-db-create-2284z" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.367172 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76s5h\" (UniqueName: \"kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h\") pod \"heat-db-create-2284z\" (UID: \"1e70b45b-b8f8-43e9-aec7-fc621835edf1\") " pod="openstack/heat-db-create-2284z" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.402810 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2284z" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.493175 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f868b9f5c-rq2jl" event={"ID":"157d6322-0be6-423b-8276-c789369b07d2","Type":"ContainerStarted","Data":"058a4937f3b6d67db8511cf9b32631ab6ab7bdc2314b4fb1e6a86878c1c61e23"} Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.493219 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f868b9f5c-rq2jl" event={"ID":"157d6322-0be6-423b-8276-c789369b07d2","Type":"ContainerStarted","Data":"f0fdd8254ca7d61c48645de1a67f6ea1e2e45a7009f71c65733fbb4df7cf7222"} Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.493229 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f868b9f5c-rq2jl" event={"ID":"157d6322-0be6-423b-8276-c789369b07d2","Type":"ContainerStarted","Data":"ded62d627b2c1a2d15b6aaa2c389a300e2c39924a3ca1fb66c35f280a70ad602"} Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.525369 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f868b9f5c-rq2jl" podStartSLOduration=2.525350882 podStartE2EDuration="2.525350882s" podCreationTimestamp="2025-10-06 10:22:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:22:56.510070371 +0000 UTC m=+6227.300095951" watchObservedRunningTime="2025-10-06 10:22:56.525350882 +0000 UTC m=+6227.315376462" Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.872650 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2284z"] Oct 06 10:22:56 crc kubenswrapper[4989]: I1006 10:22:56.936163 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:22:56 crc kubenswrapper[4989]: E1006 10:22:56.936451 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:22:57 crc kubenswrapper[4989]: I1006 10:22:57.503821 4989 generic.go:334] "Generic (PLEG): container finished" podID="1e70b45b-b8f8-43e9-aec7-fc621835edf1" containerID="25dde054c63c2797e3961c0a33dcdf13ccd390c7dd61497e77bd47cc798c7599" exitCode=0 Oct 06 10:22:57 crc kubenswrapper[4989]: I1006 10:22:57.503995 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2284z" event={"ID":"1e70b45b-b8f8-43e9-aec7-fc621835edf1","Type":"ContainerDied","Data":"25dde054c63c2797e3961c0a33dcdf13ccd390c7dd61497e77bd47cc798c7599"} Oct 06 10:22:57 crc kubenswrapper[4989]: I1006 10:22:57.504287 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2284z" event={"ID":"1e70b45b-b8f8-43e9-aec7-fc621835edf1","Type":"ContainerStarted","Data":"23b218c4781c52991b0f77d5b83981297d116cddd415b96430d199006aee731c"} Oct 06 10:22:58 crc kubenswrapper[4989]: I1006 10:22:58.899072 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2284z" Oct 06 10:22:58 crc kubenswrapper[4989]: I1006 10:22:58.998363 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76s5h\" (UniqueName: \"kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h\") pod \"1e70b45b-b8f8-43e9-aec7-fc621835edf1\" (UID: \"1e70b45b-b8f8-43e9-aec7-fc621835edf1\") " Oct 06 10:22:59 crc kubenswrapper[4989]: I1006 10:22:59.015846 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h" (OuterVolumeSpecName: "kube-api-access-76s5h") pod "1e70b45b-b8f8-43e9-aec7-fc621835edf1" (UID: "1e70b45b-b8f8-43e9-aec7-fc621835edf1"). InnerVolumeSpecName "kube-api-access-76s5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:22:59 crc kubenswrapper[4989]: I1006 10:22:59.101418 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76s5h\" (UniqueName: \"kubernetes.io/projected/1e70b45b-b8f8-43e9-aec7-fc621835edf1-kube-api-access-76s5h\") on node \"crc\" DevicePath \"\"" Oct 06 10:22:59 crc kubenswrapper[4989]: I1006 10:22:59.524458 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2284z" event={"ID":"1e70b45b-b8f8-43e9-aec7-fc621835edf1","Type":"ContainerDied","Data":"23b218c4781c52991b0f77d5b83981297d116cddd415b96430d199006aee731c"} Oct 06 10:22:59 crc kubenswrapper[4989]: I1006 10:22:59.524727 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b218c4781c52991b0f77d5b83981297d116cddd415b96430d199006aee731c" Oct 06 10:22:59 crc kubenswrapper[4989]: I1006 10:22:59.524810 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2284z" Oct 06 10:23:05 crc kubenswrapper[4989]: I1006 10:23:05.121135 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:23:05 crc kubenswrapper[4989]: I1006 10:23:05.122000 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:23:05 crc kubenswrapper[4989]: I1006 10:23:05.122590 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f868b9f5c-rq2jl" podUID="157d6322-0be6-423b-8276-c789369b07d2" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.169550 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-e57b-account-create-sfjbt"] Oct 06 10:23:06 crc kubenswrapper[4989]: E1006 10:23:06.170159 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e70b45b-b8f8-43e9-aec7-fc621835edf1" containerName="mariadb-database-create" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.170179 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e70b45b-b8f8-43e9-aec7-fc621835edf1" containerName="mariadb-database-create" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.170445 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e70b45b-b8f8-43e9-aec7-fc621835edf1" containerName="mariadb-database-create" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.171878 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.173961 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.201195 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e57b-account-create-sfjbt"] Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.350159 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvsw\" (UniqueName: \"kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw\") pod \"heat-e57b-account-create-sfjbt\" (UID: \"f1ec8cad-e2b5-4839-a98c-e3da3169daac\") " pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.452596 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvsw\" (UniqueName: \"kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw\") pod \"heat-e57b-account-create-sfjbt\" (UID: \"f1ec8cad-e2b5-4839-a98c-e3da3169daac\") " pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.475855 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvsw\" (UniqueName: \"kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw\") pod \"heat-e57b-account-create-sfjbt\" (UID: \"f1ec8cad-e2b5-4839-a98c-e3da3169daac\") " pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.495441 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:06 crc kubenswrapper[4989]: I1006 10:23:06.980525 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e57b-account-create-sfjbt"] Oct 06 10:23:07 crc kubenswrapper[4989]: I1006 10:23:07.627198 4989 generic.go:334] "Generic (PLEG): container finished" podID="f1ec8cad-e2b5-4839-a98c-e3da3169daac" containerID="39b2f69f1486873ad1cf5049bfdf8eac684e9d630f110f5dd9a25bd26badd408" exitCode=0 Oct 06 10:23:07 crc kubenswrapper[4989]: I1006 10:23:07.627813 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e57b-account-create-sfjbt" event={"ID":"f1ec8cad-e2b5-4839-a98c-e3da3169daac","Type":"ContainerDied","Data":"39b2f69f1486873ad1cf5049bfdf8eac684e9d630f110f5dd9a25bd26badd408"} Oct 06 10:23:07 crc kubenswrapper[4989]: I1006 10:23:07.629535 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e57b-account-create-sfjbt" event={"ID":"f1ec8cad-e2b5-4839-a98c-e3da3169daac","Type":"ContainerStarted","Data":"a53efad775ce372d14687cbc992a324bee92c52d2227712ad9aea60ebd11db95"} Oct 06 10:23:08 crc kubenswrapper[4989]: I1006 10:23:08.935958 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:23:08 crc kubenswrapper[4989]: E1006 10:23:08.936738 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.050676 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.210270 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjvsw\" (UniqueName: \"kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw\") pod \"f1ec8cad-e2b5-4839-a98c-e3da3169daac\" (UID: \"f1ec8cad-e2b5-4839-a98c-e3da3169daac\") " Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.231159 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw" (OuterVolumeSpecName: "kube-api-access-jjvsw") pod "f1ec8cad-e2b5-4839-a98c-e3da3169daac" (UID: "f1ec8cad-e2b5-4839-a98c-e3da3169daac"). InnerVolumeSpecName "kube-api-access-jjvsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.312770 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjvsw\" (UniqueName: \"kubernetes.io/projected/f1ec8cad-e2b5-4839-a98c-e3da3169daac-kube-api-access-jjvsw\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.656108 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e57b-account-create-sfjbt" event={"ID":"f1ec8cad-e2b5-4839-a98c-e3da3169daac","Type":"ContainerDied","Data":"a53efad775ce372d14687cbc992a324bee92c52d2227712ad9aea60ebd11db95"} Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.656499 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a53efad775ce372d14687cbc992a324bee92c52d2227712ad9aea60ebd11db95" Oct 06 10:23:09 crc kubenswrapper[4989]: I1006 10:23:09.657282 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e57b-account-create-sfjbt" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.211646 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-28p8d"] Oct 06 10:23:11 crc kubenswrapper[4989]: E1006 10:23:11.212343 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1ec8cad-e2b5-4839-a98c-e3da3169daac" containerName="mariadb-account-create" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.212358 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1ec8cad-e2b5-4839-a98c-e3da3169daac" containerName="mariadb-account-create" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.212579 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1ec8cad-e2b5-4839-a98c-e3da3169daac" containerName="mariadb-account-create" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.213245 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.216853 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-fkqxs" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.217235 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.222741 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-28p8d"] Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.362043 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.362161 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dppr7\" (UniqueName: \"kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.362248 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.464231 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.464329 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dppr7\" (UniqueName: \"kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.464396 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.472006 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.476586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.496296 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dppr7\" (UniqueName: \"kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7\") pod \"heat-db-sync-28p8d\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.534547 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:11 crc kubenswrapper[4989]: W1006 10:23:11.992429 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda261e667_22d6_4964_a078_425699474660.slice/crio-6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516 WatchSource:0}: Error finding container 6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516: Status 404 returned error can't find the container with id 6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516 Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.994548 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-28p8d"] Oct 06 10:23:11 crc kubenswrapper[4989]: I1006 10:23:11.994924 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:23:12 crc kubenswrapper[4989]: I1006 10:23:12.690149 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-28p8d" event={"ID":"a261e667-22d6-4964-a078-425699474660","Type":"ContainerStarted","Data":"6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516"} Oct 06 10:23:17 crc kubenswrapper[4989]: I1006 10:23:17.210719 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.044054 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-scfdn"] Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.054674 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-scfdn"] Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.074138 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f868b9f5c-rq2jl" Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.146411 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.146979 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon-log" containerID="cri-o://e0fbd6d1dfe04b8a18573fe65060f2861d4c2d923909672edeabe24974e944a5" gracePeriod=30 Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.147481 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" containerID="cri-o://5b6637f416ca8b50111683091d6ad8612c48384d8c4626c6d2b2fd07d6131d51" gracePeriod=30 Oct 06 10:23:19 crc kubenswrapper[4989]: I1006 10:23:19.949953 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36224a2c-ec62-4e04-8541-a09f8c2671ca" path="/var/lib/kubelet/pods/36224a2c-ec62-4e04-8541-a09f8c2671ca/volumes" Oct 06 10:23:20 crc kubenswrapper[4989]: I1006 10:23:20.777821 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-28p8d" event={"ID":"a261e667-22d6-4964-a078-425699474660","Type":"ContainerStarted","Data":"46a0d01d6b704a92bf454c1fc3c4256dab49702466c85e57a6d662691a016e60"} Oct 06 10:23:20 crc kubenswrapper[4989]: I1006 10:23:20.810212 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-28p8d" podStartSLOduration=1.764115292 podStartE2EDuration="9.810193525s" podCreationTimestamp="2025-10-06 10:23:11 +0000 UTC" firstStartedPulling="2025-10-06 10:23:11.994705153 +0000 UTC m=+6242.784730733" lastFinishedPulling="2025-10-06 10:23:20.040783356 +0000 UTC m=+6250.830808966" observedRunningTime="2025-10-06 10:23:20.803466201 +0000 UTC m=+6251.593491781" watchObservedRunningTime="2025-10-06 10:23:20.810193525 +0000 UTC m=+6251.600219105" Oct 06 10:23:21 crc kubenswrapper[4989]: I1006 10:23:21.790226 4989 generic.go:334] "Generic (PLEG): container finished" podID="a261e667-22d6-4964-a078-425699474660" containerID="46a0d01d6b704a92bf454c1fc3c4256dab49702466c85e57a6d662691a016e60" exitCode=0 Oct 06 10:23:21 crc kubenswrapper[4989]: I1006 10:23:21.790279 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-28p8d" event={"ID":"a261e667-22d6-4964-a078-425699474660","Type":"ContainerDied","Data":"46a0d01d6b704a92bf454c1fc3c4256dab49702466c85e57a6d662691a016e60"} Oct 06 10:23:21 crc kubenswrapper[4989]: I1006 10:23:21.935992 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:23:21 crc kubenswrapper[4989]: E1006 10:23:21.936395 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.510682 4989 scope.go:117] "RemoveContainer" containerID="9de599edc0650199df1987772a836f935514ddefe755bcdea1642b15903bdc60" Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.572469 4989 scope.go:117] "RemoveContainer" containerID="380b5ba3224d49a023b83048b3c208b2163059125638d24a847532795a28b5b6" Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.612147 4989 scope.go:117] "RemoveContainer" containerID="d357906a4ae8b63392d01203589675721f42273e365cd5423e87aeb0670c0423" Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.660785 4989 scope.go:117] "RemoveContainer" containerID="9ac8bc981a972f516583e07399e57dfabc4e453a08b68e32a699ab23d85c1986" Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.812295 4989 generic.go:334] "Generic (PLEG): container finished" podID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerID="5b6637f416ca8b50111683091d6ad8612c48384d8c4626c6d2b2fd07d6131d51" exitCode=0 Oct 06 10:23:22 crc kubenswrapper[4989]: I1006 10:23:22.812503 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerDied","Data":"5b6637f416ca8b50111683091d6ad8612c48384d8c4626c6d2b2fd07d6131d51"} Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.063512 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.198279 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle\") pod \"a261e667-22d6-4964-a078-425699474660\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.198399 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data\") pod \"a261e667-22d6-4964-a078-425699474660\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.198493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dppr7\" (UniqueName: \"kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7\") pod \"a261e667-22d6-4964-a078-425699474660\" (UID: \"a261e667-22d6-4964-a078-425699474660\") " Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.203582 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7" (OuterVolumeSpecName: "kube-api-access-dppr7") pod "a261e667-22d6-4964-a078-425699474660" (UID: "a261e667-22d6-4964-a078-425699474660"). InnerVolumeSpecName "kube-api-access-dppr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.224456 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a261e667-22d6-4964-a078-425699474660" (UID: "a261e667-22d6-4964-a078-425699474660"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.264297 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data" (OuterVolumeSpecName: "config-data") pod "a261e667-22d6-4964-a078-425699474660" (UID: "a261e667-22d6-4964-a078-425699474660"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.300557 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dppr7\" (UniqueName: \"kubernetes.io/projected/a261e667-22d6-4964-a078-425699474660-kube-api-access-dppr7\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.300605 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.300617 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a261e667-22d6-4964-a078-425699474660-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.826996 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-28p8d" event={"ID":"a261e667-22d6-4964-a078-425699474660","Type":"ContainerDied","Data":"6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516"} Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.827054 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e3adce81835721f802a82ce1cea6e446263dd3b4e40fc94b18c5cf73db6b516" Oct 06 10:23:23 crc kubenswrapper[4989]: I1006 10:23:23.827087 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-28p8d" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.563882 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-76b6b48b9d-lr9tc"] Oct 06 10:23:25 crc kubenswrapper[4989]: E1006 10:23:25.564937 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a261e667-22d6-4964-a078-425699474660" containerName="heat-db-sync" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.564953 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a261e667-22d6-4964-a078-425699474660" containerName="heat-db-sync" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.565229 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a261e667-22d6-4964-a078-425699474660" containerName="heat-db-sync" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.566057 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.573059 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-fkqxs" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.582556 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.582601 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.586609 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76b6b48b9d-lr9tc"] Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.744259 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5cf7458ccf-9rtfv"] Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.745532 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753375 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-combined-ca-bundle\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753425 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5nd\" (UniqueName: \"kubernetes.io/projected/9e1a109a-be53-4da4-b947-cf4184f43a0b-kube-api-access-bb5nd\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753447 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data-custom\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753473 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmkhq\" (UniqueName: \"kubernetes.io/projected/78fc292c-9988-4bd5-a61f-754743dc1729-kube-api-access-wmkhq\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753536 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753598 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data-custom\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753626 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-combined-ca-bundle\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.753665 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.755123 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.764545 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cf7458ccf-9rtfv"] Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.787711 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6fd69d8487-9nkb7"] Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.789056 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.797034 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.804442 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fd69d8487-9nkb7"] Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.928677 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.929042 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data-custom\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.932892 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-combined-ca-bundle\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.933011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.933183 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-combined-ca-bundle\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.933272 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5nd\" (UniqueName: \"kubernetes.io/projected/9e1a109a-be53-4da4-b947-cf4184f43a0b-kube-api-access-bb5nd\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.933343 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data-custom\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.933435 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmkhq\" (UniqueName: \"kubernetes.io/projected/78fc292c-9988-4bd5-a61f-754743dc1729-kube-api-access-wmkhq\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.951833 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data-custom\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.974249 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-config-data\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:25 crc kubenswrapper[4989]: I1006 10:23:25.990732 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.010560 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-config-data-custom\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.011290 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5nd\" (UniqueName: \"kubernetes.io/projected/9e1a109a-be53-4da4-b947-cf4184f43a0b-kube-api-access-bb5nd\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.014863 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e1a109a-be53-4da4-b947-cf4184f43a0b-combined-ca-bundle\") pod \"heat-cfnapi-5cf7458ccf-9rtfv\" (UID: \"9e1a109a-be53-4da4-b947-cf4184f43a0b\") " pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.015180 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fc292c-9988-4bd5-a61f-754743dc1729-combined-ca-bundle\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.015533 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmkhq\" (UniqueName: \"kubernetes.io/projected/78fc292c-9988-4bd5-a61f-754743dc1729-kube-api-access-wmkhq\") pod \"heat-engine-76b6b48b9d-lr9tc\" (UID: \"78fc292c-9988-4bd5-a61f-754743dc1729\") " pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.036894 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.037042 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrwc5\" (UniqueName: \"kubernetes.io/projected/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-kube-api-access-qrwc5\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.037092 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-combined-ca-bundle\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.037166 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data-custom\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.071009 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.144088 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data-custom\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.144286 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.144378 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrwc5\" (UniqueName: \"kubernetes.io/projected/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-kube-api-access-qrwc5\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.144435 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-combined-ca-bundle\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.161454 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data-custom\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.167562 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-config-data\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.180367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrwc5\" (UniqueName: \"kubernetes.io/projected/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-kube-api-access-qrwc5\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.182480 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb-combined-ca-bundle\") pod \"heat-api-6fd69d8487-9nkb7\" (UID: \"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb\") " pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.197050 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.451815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.676301 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cf7458ccf-9rtfv"] Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.790159 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-76b6b48b9d-lr9tc"] Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.944000 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6fd69d8487-9nkb7"] Oct 06 10:23:26 crc kubenswrapper[4989]: W1006 10:23:26.948471 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d0a7503_18c3_4f61_b6e4_7a083fb7d5fb.slice/crio-2a071aaa7d785b33e84d116f79b0c2021c530e92420c7ae650ead5f7acb3b613 WatchSource:0}: Error finding container 2a071aaa7d785b33e84d116f79b0c2021c530e92420c7ae650ead5f7acb3b613: Status 404 returned error can't find the container with id 2a071aaa7d785b33e84d116f79b0c2021c530e92420c7ae650ead5f7acb3b613 Oct 06 10:23:26 crc kubenswrapper[4989]: I1006 10:23:26.996615 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76b6b48b9d-lr9tc" event={"ID":"78fc292c-9988-4bd5-a61f-754743dc1729","Type":"ContainerStarted","Data":"dbab61e3b8666e98b4f57311ac8dfe202757467b6fd80bb9a97a3ba3cb24d599"} Oct 06 10:23:27 crc kubenswrapper[4989]: I1006 10:23:27.000514 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fd69d8487-9nkb7" event={"ID":"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb","Type":"ContainerStarted","Data":"2a071aaa7d785b33e84d116f79b0c2021c530e92420c7ae650ead5f7acb3b613"} Oct 06 10:23:27 crc kubenswrapper[4989]: I1006 10:23:27.002434 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" event={"ID":"9e1a109a-be53-4da4-b947-cf4184f43a0b","Type":"ContainerStarted","Data":"b886ec744752433d6ee225cff2bba413958099a99b4cb9060690b3d4d3241f5c"} Oct 06 10:23:27 crc kubenswrapper[4989]: I1006 10:23:27.288578 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 06 10:23:28 crc kubenswrapper[4989]: I1006 10:23:28.021951 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-76b6b48b9d-lr9tc" event={"ID":"78fc292c-9988-4bd5-a61f-754743dc1729","Type":"ContainerStarted","Data":"18ba3dd5f6ee55fd3698a242e813c74daadcdc8e27cbd632119ce0ca619f9c8d"} Oct 06 10:23:28 crc kubenswrapper[4989]: I1006 10:23:28.023025 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:28 crc kubenswrapper[4989]: I1006 10:23:28.053461 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-76b6b48b9d-lr9tc" podStartSLOduration=3.053343111 podStartE2EDuration="3.053343111s" podCreationTimestamp="2025-10-06 10:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:23:28.046409381 +0000 UTC m=+6258.836434961" watchObservedRunningTime="2025-10-06 10:23:28.053343111 +0000 UTC m=+6258.843368691" Oct 06 10:23:29 crc kubenswrapper[4989]: I1006 10:23:29.028437 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f4f3-account-create-vkrq9"] Oct 06 10:23:29 crc kubenswrapper[4989]: I1006 10:23:29.040463 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f4f3-account-create-vkrq9"] Oct 06 10:23:29 crc kubenswrapper[4989]: I1006 10:23:29.973260 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15025438-e8fe-460e-a65f-2776b76f0cbd" path="/var/lib/kubelet/pods/15025438-e8fe-460e-a65f-2776b76f0cbd/volumes" Oct 06 10:23:30 crc kubenswrapper[4989]: I1006 10:23:30.040028 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6fd69d8487-9nkb7" event={"ID":"5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb","Type":"ContainerStarted","Data":"35950ebb8af51e3359e4feabdc281989733328f51380941843e90fa8b9f82f74"} Oct 06 10:23:30 crc kubenswrapper[4989]: I1006 10:23:30.040250 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:30 crc kubenswrapper[4989]: I1006 10:23:30.046920 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" event={"ID":"9e1a109a-be53-4da4-b947-cf4184f43a0b","Type":"ContainerStarted","Data":"f81b2fd98942e668c216f594c43a9f6c01bdacd128cd8eaec7c98665a4e2357e"} Oct 06 10:23:30 crc kubenswrapper[4989]: I1006 10:23:30.047084 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:30 crc kubenswrapper[4989]: I1006 10:23:30.071071 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6fd69d8487-9nkb7" podStartSLOduration=2.957375096 podStartE2EDuration="5.071053706s" podCreationTimestamp="2025-10-06 10:23:25 +0000 UTC" firstStartedPulling="2025-10-06 10:23:26.953970561 +0000 UTC m=+6257.743996161" lastFinishedPulling="2025-10-06 10:23:29.067649191 +0000 UTC m=+6259.857674771" observedRunningTime="2025-10-06 10:23:30.058621327 +0000 UTC m=+6260.848646907" watchObservedRunningTime="2025-10-06 10:23:30.071053706 +0000 UTC m=+6260.861079296" Oct 06 10:23:34 crc kubenswrapper[4989]: I1006 10:23:34.936736 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:23:34 crc kubenswrapper[4989]: E1006 10:23:34.937540 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:23:35 crc kubenswrapper[4989]: I1006 10:23:35.034250 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" podStartSLOduration=7.658441522 podStartE2EDuration="10.034231777s" podCreationTimestamp="2025-10-06 10:23:25 +0000 UTC" firstStartedPulling="2025-10-06 10:23:26.696360585 +0000 UTC m=+6257.486386155" lastFinishedPulling="2025-10-06 10:23:29.07215083 +0000 UTC m=+6259.862176410" observedRunningTime="2025-10-06 10:23:30.083056482 +0000 UTC m=+6260.873082062" watchObservedRunningTime="2025-10-06 10:23:35.034231777 +0000 UTC m=+6265.824257357" Oct 06 10:23:35 crc kubenswrapper[4989]: I1006 10:23:35.039311 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-2vkss"] Oct 06 10:23:35 crc kubenswrapper[4989]: I1006 10:23:35.051266 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-2vkss"] Oct 06 10:23:35 crc kubenswrapper[4989]: I1006 10:23:35.977150 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e01ce706-6ac5-4f40-b472-c6ea75ff4d64" path="/var/lib/kubelet/pods/e01ce706-6ac5-4f40-b472-c6ea75ff4d64/volumes" Oct 06 10:23:37 crc kubenswrapper[4989]: I1006 10:23:37.288590 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 06 10:23:37 crc kubenswrapper[4989]: I1006 10:23:37.404116 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5cf7458ccf-9rtfv" Oct 06 10:23:38 crc kubenswrapper[4989]: I1006 10:23:37.995041 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-6fd69d8487-9nkb7" Oct 06 10:23:46 crc kubenswrapper[4989]: I1006 10:23:46.223439 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-76b6b48b9d-lr9tc" Oct 06 10:23:47 crc kubenswrapper[4989]: I1006 10:23:47.288700 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6c6ddd9d9c-kz8rv" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.109:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.109:8080: connect: connection refused" Oct 06 10:23:47 crc kubenswrapper[4989]: I1006 10:23:47.288852 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.251227 4989 generic.go:334] "Generic (PLEG): container finished" podID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerID="e0fbd6d1dfe04b8a18573fe65060f2861d4c2d923909672edeabe24974e944a5" exitCode=137 Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.251401 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerDied","Data":"e0fbd6d1dfe04b8a18573fe65060f2861d4c2d923909672edeabe24974e944a5"} Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.663379 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.819944 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data\") pod \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.820154 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key\") pod \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.820289 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs\") pod \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.820521 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5vdz\" (UniqueName: \"kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz\") pod \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.820606 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts\") pod \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\" (UID: \"d280e305-5e2d-4e2f-a421-1e0c93c598ab\") " Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.821412 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs" (OuterVolumeSpecName: "logs") pod "d280e305-5e2d-4e2f-a421-1e0c93c598ab" (UID: "d280e305-5e2d-4e2f-a421-1e0c93c598ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.829785 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d280e305-5e2d-4e2f-a421-1e0c93c598ab" (UID: "d280e305-5e2d-4e2f-a421-1e0c93c598ab"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.830286 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz" (OuterVolumeSpecName: "kube-api-access-f5vdz") pod "d280e305-5e2d-4e2f-a421-1e0c93c598ab" (UID: "d280e305-5e2d-4e2f-a421-1e0c93c598ab"). InnerVolumeSpecName "kube-api-access-f5vdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.859305 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts" (OuterVolumeSpecName: "scripts") pod "d280e305-5e2d-4e2f-a421-1e0c93c598ab" (UID: "d280e305-5e2d-4e2f-a421-1e0c93c598ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.880699 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data" (OuterVolumeSpecName: "config-data") pod "d280e305-5e2d-4e2f-a421-1e0c93c598ab" (UID: "d280e305-5e2d-4e2f-a421-1e0c93c598ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.932377 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5vdz\" (UniqueName: \"kubernetes.io/projected/d280e305-5e2d-4e2f-a421-1e0c93c598ab-kube-api-access-f5vdz\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.932412 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.932423 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d280e305-5e2d-4e2f-a421-1e0c93c598ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.932437 4989 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d280e305-5e2d-4e2f-a421-1e0c93c598ab-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.932451 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d280e305-5e2d-4e2f-a421-1e0c93c598ab-logs\") on node \"crc\" DevicePath \"\"" Oct 06 10:23:49 crc kubenswrapper[4989]: I1006 10:23:49.946065 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:23:49 crc kubenswrapper[4989]: E1006 10:23:49.946754 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.275568 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c6ddd9d9c-kz8rv" event={"ID":"d280e305-5e2d-4e2f-a421-1e0c93c598ab","Type":"ContainerDied","Data":"8f191774d71d92515dbd87ae782f83d7412ecd233fe34bf13de2677de1f045a5"} Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.275611 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c6ddd9d9c-kz8rv" Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.275733 4989 scope.go:117] "RemoveContainer" containerID="5b6637f416ca8b50111683091d6ad8612c48384d8c4626c6d2b2fd07d6131d51" Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.315511 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.332315 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c6ddd9d9c-kz8rv"] Oct 06 10:23:50 crc kubenswrapper[4989]: I1006 10:23:50.456573 4989 scope.go:117] "RemoveContainer" containerID="e0fbd6d1dfe04b8a18573fe65060f2861d4c2d923909672edeabe24974e944a5" Oct 06 10:23:51 crc kubenswrapper[4989]: I1006 10:23:51.961229 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" path="/var/lib/kubelet/pods/d280e305-5e2d-4e2f-a421-1e0c93c598ab/volumes" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.797456 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj"] Oct 06 10:24:02 crc kubenswrapper[4989]: E1006 10:24:02.798358 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.798370 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" Oct 06 10:24:02 crc kubenswrapper[4989]: E1006 10:24:02.798381 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon-log" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.798387 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon-log" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.798566 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon-log" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.798586 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d280e305-5e2d-4e2f-a421-1e0c93c598ab" containerName="horizon" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.799962 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.802146 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.815878 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj"] Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.907076 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.907181 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:02 crc kubenswrapper[4989]: I1006 10:24:02.907242 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmm2p\" (UniqueName: \"kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.009367 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.009504 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.009594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmm2p\" (UniqueName: \"kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.010003 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.010407 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.034248 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmm2p\" (UniqueName: \"kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.163295 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:03 crc kubenswrapper[4989]: I1006 10:24:03.663684 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj"] Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.433292 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerStarted","Data":"2491c02a05fc7fbb8c090a35a11456531e315dc87ebe6eeaecf81971446759d4"} Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.433711 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerStarted","Data":"08cd2d173e3139c54a0edb0e983800daabafd9b2c7b9a63126a7338ab5d2dd2b"} Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.483481 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.494246 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.494371 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.665778 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks274\" (UniqueName: \"kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.665895 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.665986 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.767473 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks274\" (UniqueName: \"kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.767566 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.767690 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.768241 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.768261 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.788697 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks274\" (UniqueName: \"kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274\") pod \"redhat-operators-mdlq8\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.836457 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:04 crc kubenswrapper[4989]: I1006 10:24:04.936331 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:24:04 crc kubenswrapper[4989]: E1006 10:24:04.938496 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:24:05 crc kubenswrapper[4989]: I1006 10:24:05.354960 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:05 crc kubenswrapper[4989]: I1006 10:24:05.442442 4989 generic.go:334] "Generic (PLEG): container finished" podID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerID="2491c02a05fc7fbb8c090a35a11456531e315dc87ebe6eeaecf81971446759d4" exitCode=0 Oct 06 10:24:05 crc kubenswrapper[4989]: I1006 10:24:05.442529 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerDied","Data":"2491c02a05fc7fbb8c090a35a11456531e315dc87ebe6eeaecf81971446759d4"} Oct 06 10:24:05 crc kubenswrapper[4989]: I1006 10:24:05.445947 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerStarted","Data":"48989b283996ba0b3325853864ef7ef6a9775d1a0c1160d0708aa2b43cd9cad1"} Oct 06 10:24:06 crc kubenswrapper[4989]: I1006 10:24:06.459138 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerID="4a7e73cb9057ffca34e5c801689ce2c0d70dd6ce461ff0868c8965319ec71e8d" exitCode=0 Oct 06 10:24:06 crc kubenswrapper[4989]: I1006 10:24:06.459235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerDied","Data":"4a7e73cb9057ffca34e5c801689ce2c0d70dd6ce461ff0868c8965319ec71e8d"} Oct 06 10:24:07 crc kubenswrapper[4989]: I1006 10:24:07.473422 4989 generic.go:334] "Generic (PLEG): container finished" podID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerID="84a8c4a955af045ab8b92da92e93da1dfc1951456239fc77b15035fb1aeebd4f" exitCode=0 Oct 06 10:24:07 crc kubenswrapper[4989]: I1006 10:24:07.473536 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerDied","Data":"84a8c4a955af045ab8b92da92e93da1dfc1951456239fc77b15035fb1aeebd4f"} Oct 06 10:24:08 crc kubenswrapper[4989]: I1006 10:24:08.487736 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerStarted","Data":"aa395d354bd69ff6fd4edb2a26810233fb0eed30d74e47f0e893b2864ad075f6"} Oct 06 10:24:08 crc kubenswrapper[4989]: I1006 10:24:08.490805 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerStarted","Data":"ed82113e129d835031d4d30b1287c5eead8b8a09afcf003d4482284b1ff6e81a"} Oct 06 10:24:08 crc kubenswrapper[4989]: I1006 10:24:08.506760 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" podStartSLOduration=5.505915888 podStartE2EDuration="6.506740429s" podCreationTimestamp="2025-10-06 10:24:02 +0000 UTC" firstStartedPulling="2025-10-06 10:24:05.445219206 +0000 UTC m=+6296.235244786" lastFinishedPulling="2025-10-06 10:24:06.446043747 +0000 UTC m=+6297.236069327" observedRunningTime="2025-10-06 10:24:08.505305168 +0000 UTC m=+6299.295330758" watchObservedRunningTime="2025-10-06 10:24:08.506740429 +0000 UTC m=+6299.296766009" Oct 06 10:24:09 crc kubenswrapper[4989]: I1006 10:24:09.513336 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerID="ed82113e129d835031d4d30b1287c5eead8b8a09afcf003d4482284b1ff6e81a" exitCode=0 Oct 06 10:24:09 crc kubenswrapper[4989]: I1006 10:24:09.513526 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerDied","Data":"ed82113e129d835031d4d30b1287c5eead8b8a09afcf003d4482284b1ff6e81a"} Oct 06 10:24:09 crc kubenswrapper[4989]: I1006 10:24:09.521096 4989 generic.go:334] "Generic (PLEG): container finished" podID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerID="aa395d354bd69ff6fd4edb2a26810233fb0eed30d74e47f0e893b2864ad075f6" exitCode=0 Oct 06 10:24:09 crc kubenswrapper[4989]: I1006 10:24:09.521155 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerDied","Data":"aa395d354bd69ff6fd4edb2a26810233fb0eed30d74e47f0e893b2864ad075f6"} Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.534954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerStarted","Data":"4eb4e3283f092b195ae7bba1a4360e0aa098f7ac79841eb0537792383206b7f1"} Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.558127 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mdlq8" podStartSLOduration=3.061246241 podStartE2EDuration="6.558106844s" podCreationTimestamp="2025-10-06 10:24:04 +0000 UTC" firstStartedPulling="2025-10-06 10:24:06.466441105 +0000 UTC m=+6297.256466675" lastFinishedPulling="2025-10-06 10:24:09.963301698 +0000 UTC m=+6300.753327278" observedRunningTime="2025-10-06 10:24:10.552866823 +0000 UTC m=+6301.342892403" watchObservedRunningTime="2025-10-06 10:24:10.558106844 +0000 UTC m=+6301.348132424" Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.901814 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.996478 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util\") pod \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.996629 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle\") pod \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.996692 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmm2p\" (UniqueName: \"kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p\") pod \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\" (UID: \"ae45c68d-94f2-4ce1-a089-eb71c94e3543\") " Oct 06 10:24:10 crc kubenswrapper[4989]: I1006 10:24:10.999605 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle" (OuterVolumeSpecName: "bundle") pod "ae45c68d-94f2-4ce1-a089-eb71c94e3543" (UID: "ae45c68d-94f2-4ce1-a089-eb71c94e3543"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.003799 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p" (OuterVolumeSpecName: "kube-api-access-bmm2p") pod "ae45c68d-94f2-4ce1-a089-eb71c94e3543" (UID: "ae45c68d-94f2-4ce1-a089-eb71c94e3543"). InnerVolumeSpecName "kube-api-access-bmm2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.006612 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util" (OuterVolumeSpecName: "util") pod "ae45c68d-94f2-4ce1-a089-eb71c94e3543" (UID: "ae45c68d-94f2-4ce1-a089-eb71c94e3543"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.098741 4989 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-util\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.098776 4989 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ae45c68d-94f2-4ce1-a089-eb71c94e3543-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.098786 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmm2p\" (UniqueName: \"kubernetes.io/projected/ae45c68d-94f2-4ce1-a089-eb71c94e3543-kube-api-access-bmm2p\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.545476 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.546017 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj" event={"ID":"ae45c68d-94f2-4ce1-a089-eb71c94e3543","Type":"ContainerDied","Data":"08cd2d173e3139c54a0edb0e983800daabafd9b2c7b9a63126a7338ab5d2dd2b"} Oct 06 10:24:11 crc kubenswrapper[4989]: I1006 10:24:11.546063 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08cd2d173e3139c54a0edb0e983800daabafd9b2c7b9a63126a7338ab5d2dd2b" Oct 06 10:24:14 crc kubenswrapper[4989]: I1006 10:24:14.837702 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:14 crc kubenswrapper[4989]: I1006 10:24:14.838156 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:15 crc kubenswrapper[4989]: I1006 10:24:15.944194 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mdlq8" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="registry-server" probeResult="failure" output=< Oct 06 10:24:15 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 10:24:15 crc kubenswrapper[4989]: > Oct 06 10:24:16 crc kubenswrapper[4989]: I1006 10:24:16.936050 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:24:16 crc kubenswrapper[4989]: E1006 10:24:16.936382 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.054117 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr"] Oct 06 10:24:20 crc kubenswrapper[4989]: E1006 10:24:20.055102 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="pull" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.055116 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="pull" Oct 06 10:24:20 crc kubenswrapper[4989]: E1006 10:24:20.055144 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="extract" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.055150 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="extract" Oct 06 10:24:20 crc kubenswrapper[4989]: E1006 10:24:20.055164 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="util" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.055170 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="util" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.055346 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae45c68d-94f2-4ce1-a089-eb71c94e3543" containerName="extract" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.056059 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.061602 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-bb96r" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.061669 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.062321 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.069662 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.161851 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.163251 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.166727 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.171295 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.177369 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-q2jzm" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.190597 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwj4j\" (UniqueName: \"kubernetes.io/projected/de27d14c-6b73-4043-a3ca-240a7e9c764f-kube-api-access-jwj4j\") pod \"obo-prometheus-operator-7c8cf85677-xmbvr\" (UID: \"de27d14c-6b73-4043-a3ca-240a7e9c764f\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.198574 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.200060 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.228001 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.293017 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.293101 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwj4j\" (UniqueName: \"kubernetes.io/projected/de27d14c-6b73-4043-a3ca-240a7e9c764f-kube-api-access-jwj4j\") pod \"obo-prometheus-operator-7c8cf85677-xmbvr\" (UID: \"de27d14c-6b73-4043-a3ca-240a7e9c764f\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.293177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.293207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.293235 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.317604 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwj4j\" (UniqueName: \"kubernetes.io/projected/de27d14c-6b73-4043-a3ca-240a7e9c764f-kube-api-access-jwj4j\") pod \"obo-prometheus-operator-7c8cf85677-xmbvr\" (UID: \"de27d14c-6b73-4043-a3ca-240a7e9c764f\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.376147 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sqmxd"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.377846 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.384392 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-zsml8" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.384942 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.395972 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sqmxd"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.397281 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.398669 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.398863 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.398969 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.399000 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.408000 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.408759 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.409922 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c082c7ef-21ff-4e43-b97c-f7b900a4a304-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7bkv2\" (UID: \"c082c7ef-21ff-4e43-b97c-f7b900a4a304\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.414304 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf294dc-d953-4dc7-afcf-939c59d64ffa-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7654b8595-7g6c9\" (UID: \"3bf294dc-d953-4dc7-afcf-939c59d64ffa\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.481497 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.500519 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxvnx\" (UniqueName: \"kubernetes.io/projected/db80ab6a-4df2-47ad-9883-83dd1e730f01-kube-api-access-wxvnx\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.500721 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/db80ab6a-4df2-47ad-9883-83dd1e730f01-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.521849 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.605083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/db80ab6a-4df2-47ad-9883-83dd1e730f01-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.605427 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxvnx\" (UniqueName: \"kubernetes.io/projected/db80ab6a-4df2-47ad-9883-83dd1e730f01-kube-api-access-wxvnx\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.612610 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/db80ab6a-4df2-47ad-9883-83dd1e730f01-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.630442 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-dfztc"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.631716 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.632766 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxvnx\" (UniqueName: \"kubernetes.io/projected/db80ab6a-4df2-47ad-9883-83dd1e730f01-kube-api-access-wxvnx\") pod \"observability-operator-cc5f78dfc-sqmxd\" (UID: \"db80ab6a-4df2-47ad-9883-83dd1e730f01\") " pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.639101 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dsh8j" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.686447 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-dfztc"] Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.706867 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/df7a3bee-462c-4a74-8449-1ea82af24bdd-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.707055 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slfjf\" (UniqueName: \"kubernetes.io/projected/df7a3bee-462c-4a74-8449-1ea82af24bdd-kube-api-access-slfjf\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.808467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/df7a3bee-462c-4a74-8449-1ea82af24bdd-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.808858 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slfjf\" (UniqueName: \"kubernetes.io/projected/df7a3bee-462c-4a74-8449-1ea82af24bdd-kube-api-access-slfjf\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.810002 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/df7a3bee-462c-4a74-8449-1ea82af24bdd-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.834532 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slfjf\" (UniqueName: \"kubernetes.io/projected/df7a3bee-462c-4a74-8449-1ea82af24bdd-kube-api-access-slfjf\") pod \"perses-operator-54bc95c9fb-dfztc\" (UID: \"df7a3bee-462c-4a74-8449-1ea82af24bdd\") " pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:20 crc kubenswrapper[4989]: I1006 10:24:20.882281 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.015447 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr"] Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.018965 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.208286 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2"] Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.330207 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9"] Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.530889 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sqmxd"] Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.669227 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-dfztc"] Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.732921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" event={"ID":"db80ab6a-4df2-47ad-9883-83dd1e730f01","Type":"ContainerStarted","Data":"ed765e2534ba06c7dd3f5b686b0cd693859d60416f271bfd7d4dc2f20dee775b"} Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.748759 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" event={"ID":"3bf294dc-d953-4dc7-afcf-939c59d64ffa","Type":"ContainerStarted","Data":"afebb80b4b5d003327228b6e2be9e2e601d6d92f17955552ed01f608bc3cf795"} Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.761938 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" event={"ID":"de27d14c-6b73-4043-a3ca-240a7e9c764f","Type":"ContainerStarted","Data":"618b2fb5926194e30a9cc6f809205b7074bbaf2f8f09c8cf94f5a4f4de84912e"} Oct 06 10:24:21 crc kubenswrapper[4989]: I1006 10:24:21.776820 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" event={"ID":"c082c7ef-21ff-4e43-b97c-f7b900a4a304","Type":"ContainerStarted","Data":"b4e862943d5c887a1c0af204d77866a352444f5cf28884d16f4d95ffd998ef44"} Oct 06 10:24:22 crc kubenswrapper[4989]: I1006 10:24:22.790389 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" event={"ID":"df7a3bee-462c-4a74-8449-1ea82af24bdd","Type":"ContainerStarted","Data":"0c9923b044d3b1d07069fca974c5a62d6de3df5fe539f80de74b0bbdbd07d9d1"} Oct 06 10:24:22 crc kubenswrapper[4989]: I1006 10:24:22.839299 4989 scope.go:117] "RemoveContainer" containerID="426bf8118a2dbdc160b9328f182934de7b9703121cbbbea313ba7f037256bc98" Oct 06 10:24:22 crc kubenswrapper[4989]: I1006 10:24:22.954362 4989 scope.go:117] "RemoveContainer" containerID="3e29d9cd39bbeec7ce629ec2784d177ff51d2c4665d5a2c5204512b3f850bab0" Oct 06 10:24:24 crc kubenswrapper[4989]: I1006 10:24:24.893914 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:24 crc kubenswrapper[4989]: I1006 10:24:24.988230 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:25 crc kubenswrapper[4989]: I1006 10:24:25.143230 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:26 crc kubenswrapper[4989]: I1006 10:24:26.837455 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mdlq8" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="registry-server" containerID="cri-o://4eb4e3283f092b195ae7bba1a4360e0aa098f7ac79841eb0537792383206b7f1" gracePeriod=2 Oct 06 10:24:27 crc kubenswrapper[4989]: I1006 10:24:27.849533 4989 generic.go:334] "Generic (PLEG): container finished" podID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerID="4eb4e3283f092b195ae7bba1a4360e0aa098f7ac79841eb0537792383206b7f1" exitCode=0 Oct 06 10:24:27 crc kubenswrapper[4989]: I1006 10:24:27.849597 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerDied","Data":"4eb4e3283f092b195ae7bba1a4360e0aa098f7ac79841eb0537792383206b7f1"} Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.817630 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.836775 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks274\" (UniqueName: \"kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274\") pod \"1dabd6a0-6412-48ee-9db1-b48b665d0135\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.836822 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content\") pod \"1dabd6a0-6412-48ee-9db1-b48b665d0135\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.836865 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities\") pod \"1dabd6a0-6412-48ee-9db1-b48b665d0135\" (UID: \"1dabd6a0-6412-48ee-9db1-b48b665d0135\") " Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.837605 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities" (OuterVolumeSpecName: "utilities") pod "1dabd6a0-6412-48ee-9db1-b48b665d0135" (UID: "1dabd6a0-6412-48ee-9db1-b48b665d0135"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.884473 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdlq8" event={"ID":"1dabd6a0-6412-48ee-9db1-b48b665d0135","Type":"ContainerDied","Data":"48989b283996ba0b3325853864ef7ef6a9775d1a0c1160d0708aa2b43cd9cad1"} Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.884520 4989 scope.go:117] "RemoveContainer" containerID="4eb4e3283f092b195ae7bba1a4360e0aa098f7ac79841eb0537792383206b7f1" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.884567 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdlq8" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.886806 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274" (OuterVolumeSpecName: "kube-api-access-ks274") pod "1dabd6a0-6412-48ee-9db1-b48b665d0135" (UID: "1dabd6a0-6412-48ee-9db1-b48b665d0135"). InnerVolumeSpecName "kube-api-access-ks274". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.915096 4989 scope.go:117] "RemoveContainer" containerID="ed82113e129d835031d4d30b1287c5eead8b8a09afcf003d4482284b1ff6e81a" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.940960 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks274\" (UniqueName: \"kubernetes.io/projected/1dabd6a0-6412-48ee-9db1-b48b665d0135-kube-api-access-ks274\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.940985 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.946639 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dabd6a0-6412-48ee-9db1-b48b665d0135" (UID: "1dabd6a0-6412-48ee-9db1-b48b665d0135"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.949671 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:24:29 crc kubenswrapper[4989]: E1006 10:24:29.950244 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:24:29 crc kubenswrapper[4989]: I1006 10:24:29.987930 4989 scope.go:117] "RemoveContainer" containerID="4a7e73cb9057ffca34e5c801689ce2c0d70dd6ce461ff0868c8965319ec71e8d" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.042788 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dabd6a0-6412-48ee-9db1-b48b665d0135-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.207533 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.218432 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mdlq8"] Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.899466 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" event={"ID":"db80ab6a-4df2-47ad-9883-83dd1e730f01","Type":"ContainerStarted","Data":"c6ef0d0afbe3d3852d6270c8c960cf835883d90bf4c05097f981e09cc6b298c8"} Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.899669 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.902121 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.902735 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" event={"ID":"3bf294dc-d953-4dc7-afcf-939c59d64ffa","Type":"ContainerStarted","Data":"05eaafc849a58fda36276e5a7692560bf0797cb2ee49d029e7a990a38e647cd5"} Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.905141 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" event={"ID":"de27d14c-6b73-4043-a3ca-240a7e9c764f","Type":"ContainerStarted","Data":"7bf6c7b060550bbd1c8a6e3cd1631504ece7f769d0c9ba1d9ff27af4f501ef22"} Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.907189 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" event={"ID":"df7a3bee-462c-4a74-8449-1ea82af24bdd","Type":"ContainerStarted","Data":"e4c9604e180c306d99b6e837aa53efc4f4b45d073ae31986847844bd2578b8ff"} Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.907668 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.909992 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" event={"ID":"c082c7ef-21ff-4e43-b97c-f7b900a4a304","Type":"ContainerStarted","Data":"112fcbb9e0a624319fbb3ff97ac19cf7ae8b7ec41f16afe895ba33fea58b3f24"} Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.925210 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-sqmxd" podStartSLOduration=2.896461209 podStartE2EDuration="10.925185649s" podCreationTimestamp="2025-10-06 10:24:20 +0000 UTC" firstStartedPulling="2025-10-06 10:24:21.562173895 +0000 UTC m=+6312.352199475" lastFinishedPulling="2025-10-06 10:24:29.590898335 +0000 UTC m=+6320.380923915" observedRunningTime="2025-10-06 10:24:30.923274983 +0000 UTC m=+6321.713300563" watchObservedRunningTime="2025-10-06 10:24:30.925185649 +0000 UTC m=+6321.715211229" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.945215 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7g6c9" podStartSLOduration=5.498563706 podStartE2EDuration="10.945194745s" podCreationTimestamp="2025-10-06 10:24:20 +0000 UTC" firstStartedPulling="2025-10-06 10:24:21.333470212 +0000 UTC m=+6312.123495792" lastFinishedPulling="2025-10-06 10:24:26.780101241 +0000 UTC m=+6317.570126831" observedRunningTime="2025-10-06 10:24:30.939905253 +0000 UTC m=+6321.729930853" watchObservedRunningTime="2025-10-06 10:24:30.945194745 +0000 UTC m=+6321.735220325" Oct 06 10:24:30 crc kubenswrapper[4989]: I1006 10:24:30.983871 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" podStartSLOduration=5.847850734 podStartE2EDuration="10.983844469s" podCreationTimestamp="2025-10-06 10:24:20 +0000 UTC" firstStartedPulling="2025-10-06 10:24:21.692994476 +0000 UTC m=+6312.483020056" lastFinishedPulling="2025-10-06 10:24:26.828988221 +0000 UTC m=+6317.619013791" observedRunningTime="2025-10-06 10:24:30.97449927 +0000 UTC m=+6321.764524850" watchObservedRunningTime="2025-10-06 10:24:30.983844469 +0000 UTC m=+6321.773870059" Oct 06 10:24:31 crc kubenswrapper[4989]: I1006 10:24:31.023008 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7654b8595-7bkv2" podStartSLOduration=5.478838889 podStartE2EDuration="11.022988448s" podCreationTimestamp="2025-10-06 10:24:20 +0000 UTC" firstStartedPulling="2025-10-06 10:24:21.234943043 +0000 UTC m=+6312.024968633" lastFinishedPulling="2025-10-06 10:24:26.779092602 +0000 UTC m=+6317.569118192" observedRunningTime="2025-10-06 10:24:30.993982082 +0000 UTC m=+6321.784007662" watchObservedRunningTime="2025-10-06 10:24:31.022988448 +0000 UTC m=+6321.813014028" Oct 06 10:24:31 crc kubenswrapper[4989]: I1006 10:24:31.063920 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-xmbvr" podStartSLOduration=5.286133543 podStartE2EDuration="11.063874217s" podCreationTimestamp="2025-10-06 10:24:20 +0000 UTC" firstStartedPulling="2025-10-06 10:24:21.050729282 +0000 UTC m=+6311.840754862" lastFinishedPulling="2025-10-06 10:24:26.828469956 +0000 UTC m=+6317.618495536" observedRunningTime="2025-10-06 10:24:31.059227453 +0000 UTC m=+6321.849253033" watchObservedRunningTime="2025-10-06 10:24:31.063874217 +0000 UTC m=+6321.853899797" Oct 06 10:24:31 crc kubenswrapper[4989]: I1006 10:24:31.946802 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" path="/var/lib/kubelet/pods/1dabd6a0-6412-48ee-9db1-b48b665d0135/volumes" Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.046145 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-68dl8"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.058188 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xxln2"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.072408 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-ls8gg"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.082795 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-68dl8"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.092261 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xxln2"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.100057 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-ls8gg"] Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.951562 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="300b916c-cba2-4226-b79c-d32d69454b49" path="/var/lib/kubelet/pods/300b916c-cba2-4226-b79c-d32d69454b49/volumes" Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.954613 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e2645a3-e847-4998-97b3-054835ab535a" path="/var/lib/kubelet/pods/4e2645a3-e847-4998-97b3-054835ab535a/volumes" Oct 06 10:24:33 crc kubenswrapper[4989]: I1006 10:24:33.955322 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5" path="/var/lib/kubelet/pods/e50c946b-ccfa-4b43-8e4f-8b47cdaaeed5/volumes" Oct 06 10:24:41 crc kubenswrapper[4989]: I1006 10:24:41.022766 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-dfztc" Oct 06 10:24:41 crc kubenswrapper[4989]: I1006 10:24:41.935561 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.029037 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46"} Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.034791 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a2dc-account-create-vjvqk"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.050094 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-40b9-account-create-7c2t2"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.060848 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-40b9-account-create-7c2t2"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.081166 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a2dc-account-create-vjvqk"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.089059 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f87a-account-create-vmjjt"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.097493 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f87a-account-create-vmjjt"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.614693 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.615445 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" containerName="openstackclient" containerID="cri-o://b658573e0eebd0b881bee255bb430656cf19dfb2d63bae4c4197cddca91f2514" gracePeriod=2 Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.632785 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.705013 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 06 10:24:43 crc kubenswrapper[4989]: E1006 10:24:43.705718 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" containerName="openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.705800 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" containerName="openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: E1006 10:24:43.705875 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="extract-content" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.705928 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="extract-content" Oct 06 10:24:43 crc kubenswrapper[4989]: E1006 10:24:43.705992 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="registry-server" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.706053 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="registry-server" Oct 06 10:24:43 crc kubenswrapper[4989]: E1006 10:24:43.706117 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="extract-utilities" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.706168 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="extract-utilities" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.706398 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" containerName="openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.706475 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dabd6a0-6412-48ee-9db1-b48b665d0135" containerName="registry-server" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.707224 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.715194 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" podUID="5cc043d8-b621-49dd-a595-d704df100397" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.736804 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.806639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc043d8-b621-49dd-a595-d704df100397-openstack-config\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.806813 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8f8l\" (UniqueName: \"kubernetes.io/projected/5cc043d8-b621-49dd-a595-d704df100397-kube-api-access-r8f8l\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.807129 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc043d8-b621-49dd-a595-d704df100397-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.912111 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8f8l\" (UniqueName: \"kubernetes.io/projected/5cc043d8-b621-49dd-a595-d704df100397-kube-api-access-r8f8l\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.912283 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc043d8-b621-49dd-a595-d704df100397-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.912356 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc043d8-b621-49dd-a595-d704df100397-openstack-config\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.913409 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5cc043d8-b621-49dd-a595-d704df100397-openstack-config\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.926466 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.928181 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.930330 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5cc043d8-b621-49dd-a595-d704df100397-openstack-config-secret\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.935061 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-tx8qc" Oct 06 10:24:43 crc kubenswrapper[4989]: I1006 10:24:43.964582 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8f8l\" (UniqueName: \"kubernetes.io/projected/5cc043d8-b621-49dd-a595-d704df100397-kube-api-access-r8f8l\") pod \"openstackclient\" (UID: \"5cc043d8-b621-49dd-a595-d704df100397\") " pod="openstack/openstackclient" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.016629 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0013c281-c9bb-4c43-b20c-4cd6cee09ebb" path="/var/lib/kubelet/pods/0013c281-c9bb-4c43-b20c-4cd6cee09ebb/volumes" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.018862 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="059c270b-9842-4228-a836-a256a05d570b" path="/var/lib/kubelet/pods/059c270b-9842-4228-a836-a256a05d570b/volumes" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.040402 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbaef9b1-96ff-4afc-a1f9-49fb7ae41636" path="/var/lib/kubelet/pods/dbaef9b1-96ff-4afc-a1f9-49fb7ae41636/volumes" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.041109 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.077214 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.146940 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdwnp\" (UniqueName: \"kubernetes.io/projected/6c6f6eab-ae0b-4cab-b028-b589004ef31a-kube-api-access-rdwnp\") pod \"kube-state-metrics-0\" (UID: \"6c6f6eab-ae0b-4cab-b028-b589004ef31a\") " pod="openstack/kube-state-metrics-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.258314 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwnp\" (UniqueName: \"kubernetes.io/projected/6c6f6eab-ae0b-4cab-b028-b589004ef31a-kube-api-access-rdwnp\") pod \"kube-state-metrics-0\" (UID: \"6c6f6eab-ae0b-4cab-b028-b589004ef31a\") " pod="openstack/kube-state-metrics-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.294752 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwnp\" (UniqueName: \"kubernetes.io/projected/6c6f6eab-ae0b-4cab-b028-b589004ef31a-kube-api-access-rdwnp\") pod \"kube-state-metrics-0\" (UID: \"6c6f6eab-ae0b-4cab-b028-b589004ef31a\") " pod="openstack/kube-state-metrics-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.427434 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.732136 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.735143 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.741804 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-5nkdg" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.742434 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.742566 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.742709 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.773819 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.885204 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt72w\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-kube-api-access-mt72w\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.885815 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.885861 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.885902 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.885958 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:44 crc kubenswrapper[4989]: I1006 10:24:44.886142 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.993798 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt72w\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-kube-api-access-mt72w\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.993860 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.993899 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.993935 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.993979 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.994044 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:44.999520 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.005934 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.008712 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e899a871-db7d-484d-88b6-8f13b07e565d-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.015996 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.017147 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e899a871-db7d-484d-88b6-8f13b07e565d-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.026965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt72w\" (UniqueName: \"kubernetes.io/projected/e899a871-db7d-484d-88b6-8f13b07e565d-kube-api-access-mt72w\") pod \"alertmanager-metric-storage-0\" (UID: \"e899a871-db7d-484d-88b6-8f13b07e565d\") " pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.119103 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.319683 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.434603 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.463293 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.465913 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471072 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471121 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471288 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471421 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471494 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.471548 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gjqf8" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640181 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640588 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640715 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf44v\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-kube-api-access-hf44v\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640766 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f5788673-b226-4b7d-a99d-dea8d93d1bed-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640822 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640859 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640881 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5788673-b226-4b7d-a99d-dea8d93d1bed-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640947 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.640979 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.745838 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f5788673-b226-4b7d-a99d-dea8d93d1bed-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.745906 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.745939 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.745959 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5788673-b226-4b7d-a99d-dea8d93d1bed-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.746025 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.746058 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.746088 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.746132 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf44v\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-kube-api-access-hf44v\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.753418 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.756440 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.757871 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f5788673-b226-4b7d-a99d-dea8d93d1bed-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.758970 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f5788673-b226-4b7d-a99d-dea8d93d1bed-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.759433 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.762065 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f5788673-b226-4b7d-a99d-dea8d93d1bed-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.771882 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf44v\" (UniqueName: \"kubernetes.io/projected/f5788673-b226-4b7d-a99d-dea8d93d1bed-kube-api-access-hf44v\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.774271 4989 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.774315 4989 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ddc23e2541385228e93e5700780f8c03fcce83c1c00de9e0305dc7e0db761a07/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:45 crc kubenswrapper[4989]: I1006 10:24:45.903848 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-852f3b3b-4228-458a-8a5e-66db76db5e13\") pod \"prometheus-metric-storage-0\" (UID: \"f5788673-b226-4b7d-a99d-dea8d93d1bed\") " pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.122136 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.131907 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6c6f6eab-ae0b-4cab-b028-b589004ef31a","Type":"ContainerStarted","Data":"a9d09ffa41b797dc62504db9d57005335734bc0b7d3135e59a23b72725816f12"} Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.139631 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.140307 4989 generic.go:334] "Generic (PLEG): container finished" podID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" containerID="b658573e0eebd0b881bee255bb430656cf19dfb2d63bae4c4197cddca91f2514" exitCode=137 Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.140472 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89cf5d902534abcbd6c28b163e19421a8002db20c4208f924622816aa70e251e" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.151330 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5cc043d8-b621-49dd-a595-d704df100397","Type":"ContainerStarted","Data":"afb0f34ee311d2519c3378b5dea0d9a81d08961a523b429ced4e90dd39b4d81a"} Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.151439 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5cc043d8-b621-49dd-a595-d704df100397","Type":"ContainerStarted","Data":"f377e95ec3ce362ca499e4cc4b97ac9cab850009ce386f04e069038006835b92"} Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.178937 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.182249 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.18222968 podStartE2EDuration="3.18222968s" podCreationTimestamp="2025-10-06 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:24:46.17323166 +0000 UTC m=+6336.963257240" watchObservedRunningTime="2025-10-06 10:24:46.18222968 +0000 UTC m=+6336.972255260" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.260050 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret\") pod \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.260397 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config\") pod \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.260452 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjxlz\" (UniqueName: \"kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz\") pod \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\" (UID: \"0cd01b39-add5-482b-9835-7aa6e59c3d8c\") " Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.273843 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz" (OuterVolumeSpecName: "kube-api-access-vjxlz") pod "0cd01b39-add5-482b-9835-7aa6e59c3d8c" (UID: "0cd01b39-add5-482b-9835-7aa6e59c3d8c"). InnerVolumeSpecName "kube-api-access-vjxlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.298031 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0cd01b39-add5-482b-9835-7aa6e59c3d8c" (UID: "0cd01b39-add5-482b-9835-7aa6e59c3d8c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.373844 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.373881 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjxlz\" (UniqueName: \"kubernetes.io/projected/0cd01b39-add5-482b-9835-7aa6e59c3d8c-kube-api-access-vjxlz\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.484009 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0cd01b39-add5-482b-9835-7aa6e59c3d8c" (UID: "0cd01b39-add5-482b-9835-7aa6e59c3d8c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.587352 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0cd01b39-add5-482b-9835-7aa6e59c3d8c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 06 10:24:46 crc kubenswrapper[4989]: I1006 10:24:46.919197 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 06 10:24:46 crc kubenswrapper[4989]: W1006 10:24:46.929602 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5788673_b226_4b7d_a99d_dea8d93d1bed.slice/crio-fa4c130b23c291c4715289e2877a6d79ab4e9e1258efa54d07290a9b88bbe63f WatchSource:0}: Error finding container fa4c130b23c291c4715289e2877a6d79ab4e9e1258efa54d07290a9b88bbe63f: Status 404 returned error can't find the container with id fa4c130b23c291c4715289e2877a6d79ab4e9e1258efa54d07290a9b88bbe63f Oct 06 10:24:47 crc kubenswrapper[4989]: I1006 10:24:47.167912 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerStarted","Data":"fa4c130b23c291c4715289e2877a6d79ab4e9e1258efa54d07290a9b88bbe63f"} Oct 06 10:24:47 crc kubenswrapper[4989]: I1006 10:24:47.175663 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e899a871-db7d-484d-88b6-8f13b07e565d","Type":"ContainerStarted","Data":"222a6e14611be068e2b482e40591febf50b43209c0f72b75ba3795568da1ceff"} Oct 06 10:24:47 crc kubenswrapper[4989]: I1006 10:24:47.175707 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 06 10:24:47 crc kubenswrapper[4989]: I1006 10:24:47.212107 4989 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" podUID="5cc043d8-b621-49dd-a595-d704df100397" Oct 06 10:24:47 crc kubenswrapper[4989]: I1006 10:24:47.954781 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cd01b39-add5-482b-9835-7aa6e59c3d8c" path="/var/lib/kubelet/pods/0cd01b39-add5-482b-9835-7aa6e59c3d8c/volumes" Oct 06 10:24:48 crc kubenswrapper[4989]: I1006 10:24:48.185477 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6c6f6eab-ae0b-4cab-b028-b589004ef31a","Type":"ContainerStarted","Data":"39763dcf27e5826a251ce9bf43cf013d186754f51ce10507f05c87ad88d00049"} Oct 06 10:24:48 crc kubenswrapper[4989]: I1006 10:24:48.185689 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 06 10:24:48 crc kubenswrapper[4989]: I1006 10:24:48.205141 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=4.670982785 podStartE2EDuration="5.205119003s" podCreationTimestamp="2025-10-06 10:24:43 +0000 UTC" firstStartedPulling="2025-10-06 10:24:45.724479194 +0000 UTC m=+6336.514504774" lastFinishedPulling="2025-10-06 10:24:46.258615412 +0000 UTC m=+6337.048640992" observedRunningTime="2025-10-06 10:24:48.198687137 +0000 UTC m=+6338.988712717" watchObservedRunningTime="2025-10-06 10:24:48.205119003 +0000 UTC m=+6338.995144583" Oct 06 10:24:53 crc kubenswrapper[4989]: I1006 10:24:53.032096 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gc6mj"] Oct 06 10:24:53 crc kubenswrapper[4989]: I1006 10:24:53.040808 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gc6mj"] Oct 06 10:24:53 crc kubenswrapper[4989]: I1006 10:24:53.239398 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerStarted","Data":"eabca67e811f6e883b74b79d0c12164a0cae2b984cff23f8b5b12d22ed0d5e37"} Oct 06 10:24:53 crc kubenswrapper[4989]: I1006 10:24:53.241700 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e899a871-db7d-484d-88b6-8f13b07e565d","Type":"ContainerStarted","Data":"6621e28e0fc31ecb52682cf29409370a08cd2f999cc82e1d2eaecfdffe5b3ce0"} Oct 06 10:24:53 crc kubenswrapper[4989]: I1006 10:24:53.948139 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3958753-2e0a-47be-b2c1-5deaa2920c85" path="/var/lib/kubelet/pods/c3958753-2e0a-47be-b2c1-5deaa2920c85/volumes" Oct 06 10:24:54 crc kubenswrapper[4989]: I1006 10:24:54.432637 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 06 10:25:00 crc kubenswrapper[4989]: I1006 10:25:00.320499 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerDied","Data":"eabca67e811f6e883b74b79d0c12164a0cae2b984cff23f8b5b12d22ed0d5e37"} Oct 06 10:25:00 crc kubenswrapper[4989]: I1006 10:25:00.320407 4989 generic.go:334] "Generic (PLEG): container finished" podID="f5788673-b226-4b7d-a99d-dea8d93d1bed" containerID="eabca67e811f6e883b74b79d0c12164a0cae2b984cff23f8b5b12d22ed0d5e37" exitCode=0 Oct 06 10:25:02 crc kubenswrapper[4989]: I1006 10:25:02.353343 4989 generic.go:334] "Generic (PLEG): container finished" podID="e899a871-db7d-484d-88b6-8f13b07e565d" containerID="6621e28e0fc31ecb52682cf29409370a08cd2f999cc82e1d2eaecfdffe5b3ce0" exitCode=0 Oct 06 10:25:02 crc kubenswrapper[4989]: I1006 10:25:02.353719 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e899a871-db7d-484d-88b6-8f13b07e565d","Type":"ContainerDied","Data":"6621e28e0fc31ecb52682cf29409370a08cd2f999cc82e1d2eaecfdffe5b3ce0"} Oct 06 10:25:05 crc kubenswrapper[4989]: I1006 10:25:05.408102 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerStarted","Data":"707fd11582fe1d6aaafcc1c99e087010c529d97c5c3cad4e9f5e229f0e3a83b6"} Oct 06 10:25:07 crc kubenswrapper[4989]: I1006 10:25:07.443435 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e899a871-db7d-484d-88b6-8f13b07e565d","Type":"ContainerStarted","Data":"3b661ca128a16a592f888a60e4934781b02a1353785afc1ad0bbb01decf30599"} Oct 06 10:25:09 crc kubenswrapper[4989]: I1006 10:25:09.472569 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerStarted","Data":"d835de5f89bafd9b3b95a5791e1ad43c96296d8d3fd0620ce6a05d77f9eea813"} Oct 06 10:25:11 crc kubenswrapper[4989]: I1006 10:25:11.047550 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nwvvc"] Oct 06 10:25:11 crc kubenswrapper[4989]: I1006 10:25:11.055626 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-nwvvc"] Oct 06 10:25:11 crc kubenswrapper[4989]: I1006 10:25:11.503928 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e899a871-db7d-484d-88b6-8f13b07e565d","Type":"ContainerStarted","Data":"ed3cd3d3147402d77524f44ee78c81c79d650553307b299f02f41df53f3cfadb"} Oct 06 10:25:11 crc kubenswrapper[4989]: I1006 10:25:11.536529 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=7.983902272 podStartE2EDuration="27.536493549s" podCreationTimestamp="2025-10-06 10:24:44 +0000 UTC" firstStartedPulling="2025-10-06 10:24:46.212183853 +0000 UTC m=+6337.002209433" lastFinishedPulling="2025-10-06 10:25:05.76477513 +0000 UTC m=+6356.554800710" observedRunningTime="2025-10-06 10:25:11.535592943 +0000 UTC m=+6362.325618533" watchObservedRunningTime="2025-10-06 10:25:11.536493549 +0000 UTC m=+6362.326519179" Oct 06 10:25:11 crc kubenswrapper[4989]: I1006 10:25:11.954546 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="349115de-5839-4178-9af2-db9765b42d09" path="/var/lib/kubelet/pods/349115de-5839-4178-9af2-db9765b42d09/volumes" Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.033214 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-66442"] Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.043105 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-66442"] Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.522901 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f5788673-b226-4b7d-a99d-dea8d93d1bed","Type":"ContainerStarted","Data":"1badfff4c7f62011815da49c8dbe5d857f40a5462136f3a38067dcb400fdaf09"} Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.524422 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.528913 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 06 10:25:12 crc kubenswrapper[4989]: I1006 10:25:12.571570 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.831893333 podStartE2EDuration="28.571544877s" podCreationTimestamp="2025-10-06 10:24:44 +0000 UTC" firstStartedPulling="2025-10-06 10:24:46.935854263 +0000 UTC m=+6337.725879843" lastFinishedPulling="2025-10-06 10:25:11.675505807 +0000 UTC m=+6362.465531387" observedRunningTime="2025-10-06 10:25:12.561031344 +0000 UTC m=+6363.351056994" watchObservedRunningTime="2025-10-06 10:25:12.571544877 +0000 UTC m=+6363.361570477" Oct 06 10:25:13 crc kubenswrapper[4989]: I1006 10:25:13.963304 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e54d59-17a2-46bc-8188-51c1d67b82d8" path="/var/lib/kubelet/pods/f7e54d59-17a2-46bc-8188-51c1d67b82d8/volumes" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.123705 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.123988 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.126644 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.565975 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.980187 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.983206 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.986770 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:25:16 crc kubenswrapper[4989]: I1006 10:25:16.989228 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.000512 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031211 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031268 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031789 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031810 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031869 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.031902 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.032163 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2bh4\" (UniqueName: \"kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170002 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170115 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2bh4\" (UniqueName: \"kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170239 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170268 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170344 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170366 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.170386 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.171731 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.172641 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.178559 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.190428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.190429 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.190918 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.195194 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2bh4\" (UniqueName: \"kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4\") pod \"ceilometer-0\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.321182 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:25:17 crc kubenswrapper[4989]: I1006 10:25:17.850485 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:25:18 crc kubenswrapper[4989]: I1006 10:25:18.587604 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerStarted","Data":"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e"} Oct 06 10:25:18 crc kubenswrapper[4989]: I1006 10:25:18.588231 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerStarted","Data":"724b41162211769d4300e9a9534fbc8e27e90fff5f468f3a1d92f212ddfbab7f"} Oct 06 10:25:19 crc kubenswrapper[4989]: I1006 10:25:19.600247 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerStarted","Data":"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949"} Oct 06 10:25:20 crc kubenswrapper[4989]: I1006 10:25:20.614206 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerStarted","Data":"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306"} Oct 06 10:25:21 crc kubenswrapper[4989]: I1006 10:25:21.625528 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerStarted","Data":"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c"} Oct 06 10:25:21 crc kubenswrapper[4989]: I1006 10:25:21.625868 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:25:21 crc kubenswrapper[4989]: I1006 10:25:21.649195 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.477374813 podStartE2EDuration="5.649176166s" podCreationTimestamp="2025-10-06 10:25:16 +0000 UTC" firstStartedPulling="2025-10-06 10:25:17.846982451 +0000 UTC m=+6368.637008041" lastFinishedPulling="2025-10-06 10:25:21.018783814 +0000 UTC m=+6371.808809394" observedRunningTime="2025-10-06 10:25:21.644317276 +0000 UTC m=+6372.434342896" watchObservedRunningTime="2025-10-06 10:25:21.649176166 +0000 UTC m=+6372.439201746" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.302547 4989 scope.go:117] "RemoveContainer" containerID="bcaac616ade48e0614a4c1777146ba0c218b9b16a0465aa989b00c7369163796" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.332676 4989 scope.go:117] "RemoveContainer" containerID="26c8ef0b6ca527dc9bafc25e08579ed1053186c26a778001e06bbd575764f951" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.418948 4989 scope.go:117] "RemoveContainer" containerID="01e69d9212f127c40568a4f89fe9b3961e9bbda94bde5151fe3355b118187f12" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.466848 4989 scope.go:117] "RemoveContainer" containerID="0f9a5aa742eb800b5c99e6eabb0e3c6458f88647a51c7255f499e8e2477c95b0" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.508108 4989 scope.go:117] "RemoveContainer" containerID="e4e9086cdf4328a23d6dbf6690f425192c84da003609ad354d32447d02120bf4" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.549349 4989 scope.go:117] "RemoveContainer" containerID="8c8739101f7578b22ef1571b78316d6e773b0f3122431cb0ca7d7ed25acdc41d" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.617249 4989 scope.go:117] "RemoveContainer" containerID="5ec1e4ec40d647079cc1dcb153c1fdee44b84ecabf63487745d1ae9c2c207227" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.666226 4989 scope.go:117] "RemoveContainer" containerID="c425c248c742879d68ba3e729e6dfd0ce90e7dfe72f4b48b20ef03ac1daf3757" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.686784 4989 scope.go:117] "RemoveContainer" containerID="90932bd0f6d74676b81c1d46a91e4164bd30a0572ad080e0142a1d6f0ad26c23" Oct 06 10:25:23 crc kubenswrapper[4989]: I1006 10:25:23.719234 4989 scope.go:117] "RemoveContainer" containerID="b658573e0eebd0b881bee255bb430656cf19dfb2d63bae4c4197cddca91f2514" Oct 06 10:25:25 crc kubenswrapper[4989]: I1006 10:25:25.946873 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-xngpw"] Oct 06 10:25:25 crc kubenswrapper[4989]: I1006 10:25:25.948682 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:25 crc kubenswrapper[4989]: I1006 10:25:25.957343 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-xngpw"] Oct 06 10:25:26 crc kubenswrapper[4989]: I1006 10:25:26.065170 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf6cz\" (UniqueName: \"kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz\") pod \"aodh-db-create-xngpw\" (UID: \"75f45596-8c3f-42fb-90db-43dfc5949438\") " pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:26 crc kubenswrapper[4989]: I1006 10:25:26.167446 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf6cz\" (UniqueName: \"kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz\") pod \"aodh-db-create-xngpw\" (UID: \"75f45596-8c3f-42fb-90db-43dfc5949438\") " pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:26 crc kubenswrapper[4989]: I1006 10:25:26.187114 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf6cz\" (UniqueName: \"kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz\") pod \"aodh-db-create-xngpw\" (UID: \"75f45596-8c3f-42fb-90db-43dfc5949438\") " pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:26 crc kubenswrapper[4989]: I1006 10:25:26.281460 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:26 crc kubenswrapper[4989]: I1006 10:25:26.878894 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-xngpw"] Oct 06 10:25:27 crc kubenswrapper[4989]: I1006 10:25:27.703819 4989 generic.go:334] "Generic (PLEG): container finished" podID="75f45596-8c3f-42fb-90db-43dfc5949438" containerID="280af0cdfc131d95d6d757b1d19dec581cee9e1b3078f04397db60079b677112" exitCode=0 Oct 06 10:25:27 crc kubenswrapper[4989]: I1006 10:25:27.703880 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-xngpw" event={"ID":"75f45596-8c3f-42fb-90db-43dfc5949438","Type":"ContainerDied","Data":"280af0cdfc131d95d6d757b1d19dec581cee9e1b3078f04397db60079b677112"} Oct 06 10:25:27 crc kubenswrapper[4989]: I1006 10:25:27.703954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-xngpw" event={"ID":"75f45596-8c3f-42fb-90db-43dfc5949438","Type":"ContainerStarted","Data":"1d32c8d6195ab47a3347d47120dd67fd774a925a2ab7ed88f5eca7c76fbfe4d1"} Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.169866 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.248671 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf6cz\" (UniqueName: \"kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz\") pod \"75f45596-8c3f-42fb-90db-43dfc5949438\" (UID: \"75f45596-8c3f-42fb-90db-43dfc5949438\") " Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.255788 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz" (OuterVolumeSpecName: "kube-api-access-pf6cz") pod "75f45596-8c3f-42fb-90db-43dfc5949438" (UID: "75f45596-8c3f-42fb-90db-43dfc5949438"). InnerVolumeSpecName "kube-api-access-pf6cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.351553 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf6cz\" (UniqueName: \"kubernetes.io/projected/75f45596-8c3f-42fb-90db-43dfc5949438-kube-api-access-pf6cz\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.737389 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-xngpw" event={"ID":"75f45596-8c3f-42fb-90db-43dfc5949438","Type":"ContainerDied","Data":"1d32c8d6195ab47a3347d47120dd67fd774a925a2ab7ed88f5eca7c76fbfe4d1"} Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.737781 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d32c8d6195ab47a3347d47120dd67fd774a925a2ab7ed88f5eca7c76fbfe4d1" Oct 06 10:25:29 crc kubenswrapper[4989]: I1006 10:25:29.737498 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-xngpw" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.044173 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-hs64s"] Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.061192 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-hs64s"] Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.772399 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:30 crc kubenswrapper[4989]: E1006 10:25:30.773687 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f45596-8c3f-42fb-90db-43dfc5949438" containerName="mariadb-database-create" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.773717 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f45596-8c3f-42fb-90db-43dfc5949438" containerName="mariadb-database-create" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.774199 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f45596-8c3f-42fb-90db-43dfc5949438" containerName="mariadb-database-create" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.776536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.797227 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.885760 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.885997 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7fk\" (UniqueName: \"kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.886096 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.988550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7fk\" (UniqueName: \"kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.988743 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.988809 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.989216 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:30 crc kubenswrapper[4989]: I1006 10:25:30.989306 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:31 crc kubenswrapper[4989]: I1006 10:25:31.012037 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7fk\" (UniqueName: \"kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk\") pod \"certified-operators-jgnsp\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:31 crc kubenswrapper[4989]: I1006 10:25:31.108301 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:31 crc kubenswrapper[4989]: I1006 10:25:31.596528 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:31 crc kubenswrapper[4989]: W1006 10:25:31.600791 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod350f3b89_fa5f_4193_b08d_0f4cb2324412.slice/crio-66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5 WatchSource:0}: Error finding container 66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5: Status 404 returned error can't find the container with id 66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5 Oct 06 10:25:31 crc kubenswrapper[4989]: I1006 10:25:31.769589 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerStarted","Data":"66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5"} Oct 06 10:25:31 crc kubenswrapper[4989]: I1006 10:25:31.948352 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76629129-a886-4f22-a326-801b77cce130" path="/var/lib/kubelet/pods/76629129-a886-4f22-a326-801b77cce130/volumes" Oct 06 10:25:32 crc kubenswrapper[4989]: I1006 10:25:32.795567 4989 generic.go:334] "Generic (PLEG): container finished" podID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerID="9e77cc225a442bd453aa46829a0fd51b93198a75e4215367499c6c3ad26c2b9e" exitCode=0 Oct 06 10:25:32 crc kubenswrapper[4989]: I1006 10:25:32.795878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerDied","Data":"9e77cc225a442bd453aa46829a0fd51b93198a75e4215367499c6c3ad26c2b9e"} Oct 06 10:25:34 crc kubenswrapper[4989]: I1006 10:25:34.831370 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerStarted","Data":"c8c182c30571bdc18d84e5b1bbf5775ad0a11f256d31e74afd05092e0476f8a1"} Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.020482 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-7710-account-create-hjtcv"] Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.023693 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.026022 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.048460 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-7710-account-create-hjtcv"] Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.108990 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpfzp\" (UniqueName: \"kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp\") pod \"aodh-7710-account-create-hjtcv\" (UID: \"3ae70841-1f24-4bc0-8b0e-3614cfa21862\") " pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.211147 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpfzp\" (UniqueName: \"kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp\") pod \"aodh-7710-account-create-hjtcv\" (UID: \"3ae70841-1f24-4bc0-8b0e-3614cfa21862\") " pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.228615 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpfzp\" (UniqueName: \"kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp\") pod \"aodh-7710-account-create-hjtcv\" (UID: \"3ae70841-1f24-4bc0-8b0e-3614cfa21862\") " pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.350995 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.830322 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-7710-account-create-hjtcv"] Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.857624 4989 generic.go:334] "Generic (PLEG): container finished" podID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerID="c8c182c30571bdc18d84e5b1bbf5775ad0a11f256d31e74afd05092e0476f8a1" exitCode=0 Oct 06 10:25:36 crc kubenswrapper[4989]: I1006 10:25:36.857685 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerDied","Data":"c8c182c30571bdc18d84e5b1bbf5775ad0a11f256d31e74afd05092e0476f8a1"} Oct 06 10:25:37 crc kubenswrapper[4989]: I1006 10:25:37.878765 4989 generic.go:334] "Generic (PLEG): container finished" podID="3ae70841-1f24-4bc0-8b0e-3614cfa21862" containerID="c46db30270c5f31c350adcc443c8f6f5ebdff1c6a4bb0d839e407143a9ef919d" exitCode=0 Oct 06 10:25:37 crc kubenswrapper[4989]: I1006 10:25:37.878910 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-7710-account-create-hjtcv" event={"ID":"3ae70841-1f24-4bc0-8b0e-3614cfa21862","Type":"ContainerDied","Data":"c46db30270c5f31c350adcc443c8f6f5ebdff1c6a4bb0d839e407143a9ef919d"} Oct 06 10:25:37 crc kubenswrapper[4989]: I1006 10:25:37.879207 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-7710-account-create-hjtcv" event={"ID":"3ae70841-1f24-4bc0-8b0e-3614cfa21862","Type":"ContainerStarted","Data":"30e7e2534c2162dd7705c0f8994cd5e03dd3745f6593a033f1f9274a04c984ee"} Oct 06 10:25:37 crc kubenswrapper[4989]: I1006 10:25:37.888609 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerStarted","Data":"1950a5dd3aec7172fd86482c62193f94c6a5254c52a32dedeed8e65ecb63ca35"} Oct 06 10:25:37 crc kubenswrapper[4989]: I1006 10:25:37.931895 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jgnsp" podStartSLOduration=3.137150668 podStartE2EDuration="7.931871882s" podCreationTimestamp="2025-10-06 10:25:30 +0000 UTC" firstStartedPulling="2025-10-06 10:25:32.798556417 +0000 UTC m=+6383.588582027" lastFinishedPulling="2025-10-06 10:25:37.593277631 +0000 UTC m=+6388.383303241" observedRunningTime="2025-10-06 10:25:37.918566428 +0000 UTC m=+6388.708592048" watchObservedRunningTime="2025-10-06 10:25:37.931871882 +0000 UTC m=+6388.721897492" Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.346474 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.478514 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpfzp\" (UniqueName: \"kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp\") pod \"3ae70841-1f24-4bc0-8b0e-3614cfa21862\" (UID: \"3ae70841-1f24-4bc0-8b0e-3614cfa21862\") " Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.486031 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp" (OuterVolumeSpecName: "kube-api-access-lpfzp") pod "3ae70841-1f24-4bc0-8b0e-3614cfa21862" (UID: "3ae70841-1f24-4bc0-8b0e-3614cfa21862"). InnerVolumeSpecName "kube-api-access-lpfzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.581285 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpfzp\" (UniqueName: \"kubernetes.io/projected/3ae70841-1f24-4bc0-8b0e-3614cfa21862-kube-api-access-lpfzp\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.912217 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-7710-account-create-hjtcv" event={"ID":"3ae70841-1f24-4bc0-8b0e-3614cfa21862","Type":"ContainerDied","Data":"30e7e2534c2162dd7705c0f8994cd5e03dd3745f6593a033f1f9274a04c984ee"} Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.912277 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30e7e2534c2162dd7705c0f8994cd5e03dd3745f6593a033f1f9274a04c984ee" Oct 06 10:25:39 crc kubenswrapper[4989]: I1006 10:25:39.912320 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-7710-account-create-hjtcv" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.109107 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.109407 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.366874 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-jj6ql"] Oct 06 10:25:41 crc kubenswrapper[4989]: E1006 10:25:41.367368 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae70841-1f24-4bc0-8b0e-3614cfa21862" containerName="mariadb-account-create" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.367393 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae70841-1f24-4bc0-8b0e-3614cfa21862" containerName="mariadb-account-create" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.367677 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae70841-1f24-4bc0-8b0e-3614cfa21862" containerName="mariadb-account-create" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.368468 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.371265 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-47qlf" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.371545 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.371772 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.385177 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jj6ql"] Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.422354 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.422454 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.422503 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txrkq\" (UniqueName: \"kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.422571 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.524754 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.524831 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.524907 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.524971 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txrkq\" (UniqueName: \"kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.531391 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.531613 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.540455 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.541869 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txrkq\" (UniqueName: \"kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq\") pod \"aodh-db-sync-jj6ql\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:41 crc kubenswrapper[4989]: I1006 10:25:41.687389 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:42 crc kubenswrapper[4989]: I1006 10:25:42.176719 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jj6ql"] Oct 06 10:25:42 crc kubenswrapper[4989]: I1006 10:25:42.178714 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jgnsp" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="registry-server" probeResult="failure" output=< Oct 06 10:25:42 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 10:25:42 crc kubenswrapper[4989]: > Oct 06 10:25:42 crc kubenswrapper[4989]: I1006 10:25:42.943687 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jj6ql" event={"ID":"e256cbde-f7d0-4065-ae22-c57ab646e143","Type":"ContainerStarted","Data":"c77921d422b64805da9f001edf6e345e9e69afd8f4a3e6f0be0a4f4529e9b5b3"} Oct 06 10:25:46 crc kubenswrapper[4989]: I1006 10:25:46.990137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jj6ql" event={"ID":"e256cbde-f7d0-4065-ae22-c57ab646e143","Type":"ContainerStarted","Data":"32e630257eacbe33dae883b0fc2e7149d94dea04439ee6f8f3f1fef7ba86bf1b"} Oct 06 10:25:47 crc kubenswrapper[4989]: I1006 10:25:47.020632 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-jj6ql" podStartSLOduration=2.423615161 podStartE2EDuration="6.020606651s" podCreationTimestamp="2025-10-06 10:25:41 +0000 UTC" firstStartedPulling="2025-10-06 10:25:42.18759564 +0000 UTC m=+6392.977621220" lastFinishedPulling="2025-10-06 10:25:45.78458713 +0000 UTC m=+6396.574612710" observedRunningTime="2025-10-06 10:25:47.010593312 +0000 UTC m=+6397.800618922" watchObservedRunningTime="2025-10-06 10:25:47.020606651 +0000 UTC m=+6397.810632261" Oct 06 10:25:47 crc kubenswrapper[4989]: I1006 10:25:47.333639 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 10:25:49 crc kubenswrapper[4989]: I1006 10:25:49.011985 4989 generic.go:334] "Generic (PLEG): container finished" podID="e256cbde-f7d0-4065-ae22-c57ab646e143" containerID="32e630257eacbe33dae883b0fc2e7149d94dea04439ee6f8f3f1fef7ba86bf1b" exitCode=0 Oct 06 10:25:49 crc kubenswrapper[4989]: I1006 10:25:49.012073 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jj6ql" event={"ID":"e256cbde-f7d0-4065-ae22-c57ab646e143","Type":"ContainerDied","Data":"32e630257eacbe33dae883b0fc2e7149d94dea04439ee6f8f3f1fef7ba86bf1b"} Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.466403 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.534775 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data\") pod \"e256cbde-f7d0-4065-ae22-c57ab646e143\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.534856 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts\") pod \"e256cbde-f7d0-4065-ae22-c57ab646e143\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.535053 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle\") pod \"e256cbde-f7d0-4065-ae22-c57ab646e143\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.535180 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txrkq\" (UniqueName: \"kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq\") pod \"e256cbde-f7d0-4065-ae22-c57ab646e143\" (UID: \"e256cbde-f7d0-4065-ae22-c57ab646e143\") " Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.541027 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts" (OuterVolumeSpecName: "scripts") pod "e256cbde-f7d0-4065-ae22-c57ab646e143" (UID: "e256cbde-f7d0-4065-ae22-c57ab646e143"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.541815 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq" (OuterVolumeSpecName: "kube-api-access-txrkq") pod "e256cbde-f7d0-4065-ae22-c57ab646e143" (UID: "e256cbde-f7d0-4065-ae22-c57ab646e143"). InnerVolumeSpecName "kube-api-access-txrkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.567513 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e256cbde-f7d0-4065-ae22-c57ab646e143" (UID: "e256cbde-f7d0-4065-ae22-c57ab646e143"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.571938 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data" (OuterVolumeSpecName: "config-data") pod "e256cbde-f7d0-4065-ae22-c57ab646e143" (UID: "e256cbde-f7d0-4065-ae22-c57ab646e143"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.638534 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.638567 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.638579 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e256cbde-f7d0-4065-ae22-c57ab646e143-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:50 crc kubenswrapper[4989]: I1006 10:25:50.638590 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txrkq\" (UniqueName: \"kubernetes.io/projected/e256cbde-f7d0-4065-ae22-c57ab646e143-kube-api-access-txrkq\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.034215 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jj6ql" event={"ID":"e256cbde-f7d0-4065-ae22-c57ab646e143","Type":"ContainerDied","Data":"c77921d422b64805da9f001edf6e345e9e69afd8f4a3e6f0be0a4f4529e9b5b3"} Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.034556 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c77921d422b64805da9f001edf6e345e9e69afd8f4a3e6f0be0a4f4529e9b5b3" Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.034617 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jj6ql" Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.196160 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.263183 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:51 crc kubenswrapper[4989]: I1006 10:25:51.433046 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:53 crc kubenswrapper[4989]: I1006 10:25:53.061484 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jgnsp" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="registry-server" containerID="cri-o://1950a5dd3aec7172fd86482c62193f94c6a5254c52a32dedeed8e65ecb63ca35" gracePeriod=2 Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.078223 4989 generic.go:334] "Generic (PLEG): container finished" podID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerID="1950a5dd3aec7172fd86482c62193f94c6a5254c52a32dedeed8e65ecb63ca35" exitCode=0 Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.078294 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerDied","Data":"1950a5dd3aec7172fd86482c62193f94c6a5254c52a32dedeed8e65ecb63ca35"} Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.078917 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jgnsp" event={"ID":"350f3b89-fa5f-4193-b08d-0f4cb2324412","Type":"ContainerDied","Data":"66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5"} Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.078944 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66d1b5d6a47899ff08147c5ef607ba538344d0f4692e0823203e9f9aa2307da5" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.172119 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.218493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk7fk\" (UniqueName: \"kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk\") pod \"350f3b89-fa5f-4193-b08d-0f4cb2324412\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.218714 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content\") pod \"350f3b89-fa5f-4193-b08d-0f4cb2324412\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.218744 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities\") pod \"350f3b89-fa5f-4193-b08d-0f4cb2324412\" (UID: \"350f3b89-fa5f-4193-b08d-0f4cb2324412\") " Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.220167 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities" (OuterVolumeSpecName: "utilities") pod "350f3b89-fa5f-4193-b08d-0f4cb2324412" (UID: "350f3b89-fa5f-4193-b08d-0f4cb2324412"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.224754 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk" (OuterVolumeSpecName: "kube-api-access-xk7fk") pod "350f3b89-fa5f-4193-b08d-0f4cb2324412" (UID: "350f3b89-fa5f-4193-b08d-0f4cb2324412"). InnerVolumeSpecName "kube-api-access-xk7fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.262437 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "350f3b89-fa5f-4193-b08d-0f4cb2324412" (UID: "350f3b89-fa5f-4193-b08d-0f4cb2324412"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.321282 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.321320 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350f3b89-fa5f-4193-b08d-0f4cb2324412-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:54 crc kubenswrapper[4989]: I1006 10:25:54.321342 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk7fk\" (UniqueName: \"kubernetes.io/projected/350f3b89-fa5f-4193-b08d-0f4cb2324412-kube-api-access-xk7fk\") on node \"crc\" DevicePath \"\"" Oct 06 10:25:55 crc kubenswrapper[4989]: I1006 10:25:55.092044 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jgnsp" Oct 06 10:25:55 crc kubenswrapper[4989]: I1006 10:25:55.158252 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:55 crc kubenswrapper[4989]: I1006 10:25:55.169289 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jgnsp"] Oct 06 10:25:55 crc kubenswrapper[4989]: I1006 10:25:55.953468 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" path="/var/lib/kubelet/pods/350f3b89-fa5f-4193-b08d-0f4cb2324412/volumes" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.126841 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 06 10:25:56 crc kubenswrapper[4989]: E1006 10:25:56.138007 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e256cbde-f7d0-4065-ae22-c57ab646e143" containerName="aodh-db-sync" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.138056 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e256cbde-f7d0-4065-ae22-c57ab646e143" containerName="aodh-db-sync" Oct 06 10:25:56 crc kubenswrapper[4989]: E1006 10:25:56.138098 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="registry-server" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.138105 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="registry-server" Oct 06 10:25:56 crc kubenswrapper[4989]: E1006 10:25:56.138123 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="extract-utilities" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.138129 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="extract-utilities" Oct 06 10:25:56 crc kubenswrapper[4989]: E1006 10:25:56.138141 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="extract-content" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.138150 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="extract-content" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.141559 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="350f3b89-fa5f-4193-b08d-0f4cb2324412" containerName="registry-server" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.141623 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e256cbde-f7d0-4065-ae22-c57ab646e143" containerName="aodh-db-sync" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.171559 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.181903 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.185821 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-47qlf" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.186084 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.231211 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.274419 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrd8h\" (UniqueName: \"kubernetes.io/projected/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-kube-api-access-jrd8h\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.274493 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-scripts\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.274530 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-config-data\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.274558 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.376724 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.376900 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrd8h\" (UniqueName: \"kubernetes.io/projected/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-kube-api-access-jrd8h\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.377691 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-scripts\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.377740 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-config-data\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.385492 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-scripts\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.385715 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.386233 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-config-data\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.396328 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrd8h\" (UniqueName: \"kubernetes.io/projected/b89c8d47-43dd-4d8a-b31f-a7e5655d0427-kube-api-access-jrd8h\") pod \"aodh-0\" (UID: \"b89c8d47-43dd-4d8a-b31f-a7e5655d0427\") " pod="openstack/aodh-0" Oct 06 10:25:56 crc kubenswrapper[4989]: I1006 10:25:56.503817 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 06 10:25:57 crc kubenswrapper[4989]: I1006 10:25:57.022422 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 06 10:25:57 crc kubenswrapper[4989]: I1006 10:25:57.116618 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b89c8d47-43dd-4d8a-b31f-a7e5655d0427","Type":"ContainerStarted","Data":"84c979c921549d603fced8fa0421ca7c52e76d6d734015df9f0f1f1030bebc77"} Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.128265 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b89c8d47-43dd-4d8a-b31f-a7e5655d0427","Type":"ContainerStarted","Data":"946c97502bfe7e6cc702a08e5e6a73d703dde9af121441f69e9ad98dc5f93b67"} Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.290873 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.291361 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-central-agent" containerID="cri-o://88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e" gracePeriod=30 Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.292399 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="proxy-httpd" containerID="cri-o://9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c" gracePeriod=30 Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.292531 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="sg-core" containerID="cri-o://34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306" gracePeriod=30 Oct 06 10:25:58 crc kubenswrapper[4989]: I1006 10:25:58.292607 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-notification-agent" containerID="cri-o://5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949" gracePeriod=30 Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140012 4989 generic.go:334] "Generic (PLEG): container finished" podID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerID="9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c" exitCode=0 Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140503 4989 generic.go:334] "Generic (PLEG): container finished" podID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerID="34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306" exitCode=2 Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140513 4989 generic.go:334] "Generic (PLEG): container finished" podID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerID="88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e" exitCode=0 Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140097 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerDied","Data":"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c"} Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140548 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerDied","Data":"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306"} Oct 06 10:25:59 crc kubenswrapper[4989]: I1006 10:25:59.140560 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerDied","Data":"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e"} Oct 06 10:26:00 crc kubenswrapper[4989]: I1006 10:26:00.153909 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b89c8d47-43dd-4d8a-b31f-a7e5655d0427","Type":"ContainerStarted","Data":"1b19e38df49da4e9230d60daa188bfdd3e5cf729c53d3c23c7521c4db7e54759"} Oct 06 10:26:01 crc kubenswrapper[4989]: I1006 10:26:01.164849 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b89c8d47-43dd-4d8a-b31f-a7e5655d0427","Type":"ContainerStarted","Data":"ec942e4aceb6824f48c6a04dd04ba6e08b8671ab76d9efc72c56b2aabd37997f"} Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.171895 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.178609 4989 generic.go:334] "Generic (PLEG): container finished" podID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerID="5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949" exitCode=0 Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.178975 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerDied","Data":"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949"} Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.179034 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db7a2a3d-6872-4893-9d7b-e60cba0f8844","Type":"ContainerDied","Data":"724b41162211769d4300e9a9534fbc8e27e90fff5f468f3a1d92f212ddfbab7f"} Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.179063 4989 scope.go:117] "RemoveContainer" containerID="9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.179244 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.182442 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b89c8d47-43dd-4d8a-b31f-a7e5655d0427","Type":"ContainerStarted","Data":"23f693c2122f2c4e9ff990ab31f57a7d523abba1deb98036251ea5d399b038ea"} Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.204574 4989 scope.go:117] "RemoveContainer" containerID="34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.236960 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.523022602 podStartE2EDuration="6.236940648s" podCreationTimestamp="2025-10-06 10:25:56 +0000 UTC" firstStartedPulling="2025-10-06 10:25:57.017808828 +0000 UTC m=+6407.807834408" lastFinishedPulling="2025-10-06 10:26:01.731726874 +0000 UTC m=+6412.521752454" observedRunningTime="2025-10-06 10:26:02.231095879 +0000 UTC m=+6413.021121479" watchObservedRunningTime="2025-10-06 10:26:02.236940648 +0000 UTC m=+6413.026966228" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.247117 4989 scope.go:117] "RemoveContainer" containerID="5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.293805 4989 scope.go:117] "RemoveContainer" containerID="88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.311479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2bh4\" (UniqueName: \"kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312168 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312297 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312345 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312469 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312490 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312608 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle\") pod \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\" (UID: \"db7a2a3d-6872-4893-9d7b-e60cba0f8844\") " Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.312634 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.313025 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.313431 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.313464 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db7a2a3d-6872-4893-9d7b-e60cba0f8844-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.318263 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts" (OuterVolumeSpecName: "scripts") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.322102 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4" (OuterVolumeSpecName: "kube-api-access-s2bh4") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "kube-api-access-s2bh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.344186 4989 scope.go:117] "RemoveContainer" containerID="9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.344610 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c\": container with ID starting with 9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c not found: ID does not exist" containerID="9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.344667 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c"} err="failed to get container status \"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c\": rpc error: code = NotFound desc = could not find container \"9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c\": container with ID starting with 9d7afce567a714f52871bc143772c16dca3ea8472d451f5a2d65a9a23777ef1c not found: ID does not exist" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.344687 4989 scope.go:117] "RemoveContainer" containerID="34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.345636 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306\": container with ID starting with 34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306 not found: ID does not exist" containerID="34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.345684 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306"} err="failed to get container status \"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306\": rpc error: code = NotFound desc = could not find container \"34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306\": container with ID starting with 34fd440f33e6fc7f6931ba754d9ec2857f581964caa81e10782b772e26cf1306 not found: ID does not exist" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.345699 4989 scope.go:117] "RemoveContainer" containerID="5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.345954 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949\": container with ID starting with 5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949 not found: ID does not exist" containerID="5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.345992 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949"} err="failed to get container status \"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949\": rpc error: code = NotFound desc = could not find container \"5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949\": container with ID starting with 5cc01bf2c2c435dca2d1b19358ae85ff5a4b64f97304509b7cf5aa441bb3c949 not found: ID does not exist" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.346005 4989 scope.go:117] "RemoveContainer" containerID="88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.357343 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e\": container with ID starting with 88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e not found: ID does not exist" containerID="88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.357398 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e"} err="failed to get container status \"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e\": rpc error: code = NotFound desc = could not find container \"88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e\": container with ID starting with 88b000875e607de0513597a9950557b25f5ffc64156bc4327fccf5f7c8c5e02e not found: ID does not exist" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.375843 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.409942 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.416610 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.416632 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.416641 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.416660 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2bh4\" (UniqueName: \"kubernetes.io/projected/db7a2a3d-6872-4893-9d7b-e60cba0f8844-kube-api-access-s2bh4\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.478920 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data" (OuterVolumeSpecName: "config-data") pod "db7a2a3d-6872-4893-9d7b-e60cba0f8844" (UID: "db7a2a3d-6872-4893-9d7b-e60cba0f8844"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.519347 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db7a2a3d-6872-4893-9d7b-e60cba0f8844-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.531722 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.552008 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.569559 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.570052 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-notification-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570071 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-notification-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.570082 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-central-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570088 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-central-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.570126 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="sg-core" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570133 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="sg-core" Oct 06 10:26:02 crc kubenswrapper[4989]: E1006 10:26:02.570151 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="proxy-httpd" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570157 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="proxy-httpd" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570363 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="sg-core" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570383 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-central-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570423 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="proxy-httpd" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.570439 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" containerName="ceilometer-notification-agent" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.574166 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.580751 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.580947 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.585220 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621191 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621364 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qxxf\" (UniqueName: \"kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621431 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621576 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621743 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.621799 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.723798 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.723934 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qxxf\" (UniqueName: \"kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.723999 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.724035 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.724092 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.724121 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.724143 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.724609 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.725566 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.728401 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.728489 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.738436 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.738899 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.751554 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qxxf\" (UniqueName: \"kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf\") pod \"ceilometer-0\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " pod="openstack/ceilometer-0" Oct 06 10:26:02 crc kubenswrapper[4989]: I1006 10:26:02.909016 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:03 crc kubenswrapper[4989]: I1006 10:26:03.420868 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:03 crc kubenswrapper[4989]: I1006 10:26:03.949908 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db7a2a3d-6872-4893-9d7b-e60cba0f8844" path="/var/lib/kubelet/pods/db7a2a3d-6872-4893-9d7b-e60cba0f8844/volumes" Oct 06 10:26:04 crc kubenswrapper[4989]: I1006 10:26:04.210277 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerStarted","Data":"e9478c871c99b639490e7dc6017f55105374bc3f4e965f914cc089a418dd3cf9"} Oct 06 10:26:04 crc kubenswrapper[4989]: I1006 10:26:04.210580 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerStarted","Data":"89c0d02f520206654e5c4058a2e47022e7eb4890f42565eda84b1905af0d1bbc"} Oct 06 10:26:05 crc kubenswrapper[4989]: I1006 10:26:05.230103 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerStarted","Data":"b4d5a7a243e105debe37a93dda262c322d01317642df4fd135ae924082b79b0f"} Oct 06 10:26:06 crc kubenswrapper[4989]: I1006 10:26:06.242490 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerStarted","Data":"48911551e646ebe3307d957ccdd3f287fdf2aa3db598a31e2c52563cf464a005"} Oct 06 10:26:08 crc kubenswrapper[4989]: I1006 10:26:08.288556 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerStarted","Data":"550e55d51c1abe9cd0b737bd44613026589ccf004d3fc2c5ca975e3d397de4b8"} Oct 06 10:26:08 crc kubenswrapper[4989]: I1006 10:26:08.289054 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:26:08 crc kubenswrapper[4989]: I1006 10:26:08.327015 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5381416850000003 podStartE2EDuration="6.326992025s" podCreationTimestamp="2025-10-06 10:26:02 +0000 UTC" firstStartedPulling="2025-10-06 10:26:03.433426049 +0000 UTC m=+6414.223451629" lastFinishedPulling="2025-10-06 10:26:07.222276349 +0000 UTC m=+6418.012301969" observedRunningTime="2025-10-06 10:26:08.318384307 +0000 UTC m=+6419.108409887" watchObservedRunningTime="2025-10-06 10:26:08.326992025 +0000 UTC m=+6419.117017615" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.060391 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-449sh"] Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.062404 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-449sh" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.075488 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-449sh"] Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.091333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbd2f\" (UniqueName: \"kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f\") pod \"manila-db-create-449sh\" (UID: \"2eac29ac-df99-4009-9847-6c38dd9da407\") " pod="openstack/manila-db-create-449sh" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.193394 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbd2f\" (UniqueName: \"kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f\") pod \"manila-db-create-449sh\" (UID: \"2eac29ac-df99-4009-9847-6c38dd9da407\") " pod="openstack/manila-db-create-449sh" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.216377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbd2f\" (UniqueName: \"kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f\") pod \"manila-db-create-449sh\" (UID: \"2eac29ac-df99-4009-9847-6c38dd9da407\") " pod="openstack/manila-db-create-449sh" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.398378 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-449sh" Oct 06 10:26:09 crc kubenswrapper[4989]: I1006 10:26:09.957128 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-449sh"] Oct 06 10:26:09 crc kubenswrapper[4989]: W1006 10:26:09.968501 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2eac29ac_df99_4009_9847_6c38dd9da407.slice/crio-425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e WatchSource:0}: Error finding container 425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e: Status 404 returned error can't find the container with id 425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e Oct 06 10:26:10 crc kubenswrapper[4989]: I1006 10:26:10.322008 4989 generic.go:334] "Generic (PLEG): container finished" podID="2eac29ac-df99-4009-9847-6c38dd9da407" containerID="6a947d5bead38a90aca3a239cae3d9cdd91335945936f0d8c91dd6926b1bbcb9" exitCode=0 Oct 06 10:26:10 crc kubenswrapper[4989]: I1006 10:26:10.322512 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-449sh" event={"ID":"2eac29ac-df99-4009-9847-6c38dd9da407","Type":"ContainerDied","Data":"6a947d5bead38a90aca3a239cae3d9cdd91335945936f0d8c91dd6926b1bbcb9"} Oct 06 10:26:10 crc kubenswrapper[4989]: I1006 10:26:10.322543 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-449sh" event={"ID":"2eac29ac-df99-4009-9847-6c38dd9da407","Type":"ContainerStarted","Data":"425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e"} Oct 06 10:26:11 crc kubenswrapper[4989]: I1006 10:26:11.845755 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-449sh" Oct 06 10:26:11 crc kubenswrapper[4989]: I1006 10:26:11.956096 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbd2f\" (UniqueName: \"kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f\") pod \"2eac29ac-df99-4009-9847-6c38dd9da407\" (UID: \"2eac29ac-df99-4009-9847-6c38dd9da407\") " Oct 06 10:26:11 crc kubenswrapper[4989]: I1006 10:26:11.964712 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f" (OuterVolumeSpecName: "kube-api-access-cbd2f") pod "2eac29ac-df99-4009-9847-6c38dd9da407" (UID: "2eac29ac-df99-4009-9847-6c38dd9da407"). InnerVolumeSpecName "kube-api-access-cbd2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:12 crc kubenswrapper[4989]: I1006 10:26:12.060812 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbd2f\" (UniqueName: \"kubernetes.io/projected/2eac29ac-df99-4009-9847-6c38dd9da407-kube-api-access-cbd2f\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:12 crc kubenswrapper[4989]: I1006 10:26:12.365476 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-449sh" event={"ID":"2eac29ac-df99-4009-9847-6c38dd9da407","Type":"ContainerDied","Data":"425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e"} Oct 06 10:26:12 crc kubenswrapper[4989]: I1006 10:26:12.365756 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="425dc1adf9f32050624872a6c9f0d52dfec0a6e5e30da2b2aaf110a4fa5c0c4e" Oct 06 10:26:12 crc kubenswrapper[4989]: I1006 10:26:12.365698 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-449sh" Oct 06 10:26:13 crc kubenswrapper[4989]: I1006 10:26:13.050156 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bl89j"] Oct 06 10:26:13 crc kubenswrapper[4989]: I1006 10:26:13.063023 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bl89j"] Oct 06 10:26:13 crc kubenswrapper[4989]: I1006 10:26:13.953145 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65adde8-6511-4eae-8bc5-0ca4980761fb" path="/var/lib/kubelet/pods/d65adde8-6511-4eae-8bc5-0ca4980761fb/volumes" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.160701 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-fe32-account-create-kgsdv"] Oct 06 10:26:19 crc kubenswrapper[4989]: E1006 10:26:19.161932 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eac29ac-df99-4009-9847-6c38dd9da407" containerName="mariadb-database-create" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.161949 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eac29ac-df99-4009-9847-6c38dd9da407" containerName="mariadb-database-create" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.162276 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eac29ac-df99-4009-9847-6c38dd9da407" containerName="mariadb-database-create" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.163400 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.165988 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.173722 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-fe32-account-create-kgsdv"] Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.313019 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5j2j\" (UniqueName: \"kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j\") pod \"manila-fe32-account-create-kgsdv\" (UID: \"7f6a5693-a77c-4f1b-81b9-643a410095cf\") " pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.415948 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5j2j\" (UniqueName: \"kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j\") pod \"manila-fe32-account-create-kgsdv\" (UID: \"7f6a5693-a77c-4f1b-81b9-643a410095cf\") " pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.440838 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5j2j\" (UniqueName: \"kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j\") pod \"manila-fe32-account-create-kgsdv\" (UID: \"7f6a5693-a77c-4f1b-81b9-643a410095cf\") " pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.485498 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:19 crc kubenswrapper[4989]: I1006 10:26:19.997639 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-fe32-account-create-kgsdv"] Oct 06 10:26:20 crc kubenswrapper[4989]: I1006 10:26:20.452406 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f6a5693-a77c-4f1b-81b9-643a410095cf" containerID="0983db184f561963ee77c819139824b8befeae52e528a4e79f24cf473261a92d" exitCode=0 Oct 06 10:26:20 crc kubenswrapper[4989]: I1006 10:26:20.452477 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe32-account-create-kgsdv" event={"ID":"7f6a5693-a77c-4f1b-81b9-643a410095cf","Type":"ContainerDied","Data":"0983db184f561963ee77c819139824b8befeae52e528a4e79f24cf473261a92d"} Oct 06 10:26:20 crc kubenswrapper[4989]: I1006 10:26:20.452740 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe32-account-create-kgsdv" event={"ID":"7f6a5693-a77c-4f1b-81b9-643a410095cf","Type":"ContainerStarted","Data":"4d7ea8f2d19bbc5148f744439e78030c769d389c8d66ed4d7a0fa0c0fe41ca87"} Oct 06 10:26:21 crc kubenswrapper[4989]: I1006 10:26:21.879271 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.070259 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5j2j\" (UniqueName: \"kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j\") pod \"7f6a5693-a77c-4f1b-81b9-643a410095cf\" (UID: \"7f6a5693-a77c-4f1b-81b9-643a410095cf\") " Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.076488 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j" (OuterVolumeSpecName: "kube-api-access-c5j2j") pod "7f6a5693-a77c-4f1b-81b9-643a410095cf" (UID: "7f6a5693-a77c-4f1b-81b9-643a410095cf"). InnerVolumeSpecName "kube-api-access-c5j2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.174374 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5j2j\" (UniqueName: \"kubernetes.io/projected/7f6a5693-a77c-4f1b-81b9-643a410095cf-kube-api-access-c5j2j\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.477047 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-fe32-account-create-kgsdv" event={"ID":"7f6a5693-a77c-4f1b-81b9-643a410095cf","Type":"ContainerDied","Data":"4d7ea8f2d19bbc5148f744439e78030c769d389c8d66ed4d7a0fa0c0fe41ca87"} Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.477092 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d7ea8f2d19bbc5148f744439e78030c769d389c8d66ed4d7a0fa0c0fe41ca87" Oct 06 10:26:22 crc kubenswrapper[4989]: I1006 10:26:22.477139 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-fe32-account-create-kgsdv" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.027203 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-079f-account-create-jj5g2"] Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.032382 4989 scope.go:117] "RemoveContainer" containerID="bc5254157387b769a9aea3536383115a1d98305236839efbbdcfd48b8a7170f9" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.040785 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-079f-account-create-jj5g2"] Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.070354 4989 scope.go:117] "RemoveContainer" containerID="8449a31a37b17ebae2efa26a1e1357e2d60608edecd6049c6e36c36856614fcd" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.147059 4989 scope.go:117] "RemoveContainer" containerID="85733463780b84027e139388a1aa6544e9a1e472a5326431061b545fed7e66a6" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.626691 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-7tsmz"] Oct 06 10:26:24 crc kubenswrapper[4989]: E1006 10:26:24.627530 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f6a5693-a77c-4f1b-81b9-643a410095cf" containerName="mariadb-account-create" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.627552 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f6a5693-a77c-4f1b-81b9-643a410095cf" containerName="mariadb-account-create" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.627917 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f6a5693-a77c-4f1b-81b9-643a410095cf" containerName="mariadb-account-create" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.628835 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.630985 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.630985 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-2jwdh" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.637873 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-7tsmz"] Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.741983 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.742048 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.742697 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.742847 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gfhb\" (UniqueName: \"kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.845540 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gfhb\" (UniqueName: \"kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.845771 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.845808 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.845886 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.852418 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.852472 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.859014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.867608 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gfhb\" (UniqueName: \"kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb\") pod \"manila-db-sync-7tsmz\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:24 crc kubenswrapper[4989]: I1006 10:26:24.952102 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:25 crc kubenswrapper[4989]: I1006 10:26:25.604784 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-7tsmz"] Oct 06 10:26:25 crc kubenswrapper[4989]: W1006 10:26:25.609025 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2be20a5_53e0_4fc6_a4d0_29206fbf5676.slice/crio-d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98 WatchSource:0}: Error finding container d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98: Status 404 returned error can't find the container with id d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98 Oct 06 10:26:25 crc kubenswrapper[4989]: I1006 10:26:25.955007 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e46a0615-6b06-48e6-9769-43533448ae24" path="/var/lib/kubelet/pods/e46a0615-6b06-48e6-9769-43533448ae24/volumes" Oct 06 10:26:26 crc kubenswrapper[4989]: I1006 10:26:26.533895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7tsmz" event={"ID":"b2be20a5-53e0-4fc6-a4d0-29206fbf5676","Type":"ContainerStarted","Data":"d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98"} Oct 06 10:26:31 crc kubenswrapper[4989]: I1006 10:26:31.035010 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-j76hm"] Oct 06 10:26:31 crc kubenswrapper[4989]: I1006 10:26:31.043953 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-j76hm"] Oct 06 10:26:31 crc kubenswrapper[4989]: I1006 10:26:31.594530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7tsmz" event={"ID":"b2be20a5-53e0-4fc6-a4d0-29206fbf5676","Type":"ContainerStarted","Data":"51e8997cfb34cbc8614a653d46de4ece77c1dd218cdaf51c108214506a3c35c8"} Oct 06 10:26:31 crc kubenswrapper[4989]: I1006 10:26:31.622605 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-7tsmz" podStartSLOduration=2.628852167 podStartE2EDuration="7.62257845s" podCreationTimestamp="2025-10-06 10:26:24 +0000 UTC" firstStartedPulling="2025-10-06 10:26:25.611880591 +0000 UTC m=+6436.401906191" lastFinishedPulling="2025-10-06 10:26:30.605606894 +0000 UTC m=+6441.395632474" observedRunningTime="2025-10-06 10:26:31.615749393 +0000 UTC m=+6442.405775003" watchObservedRunningTime="2025-10-06 10:26:31.62257845 +0000 UTC m=+6442.412604080" Oct 06 10:26:31 crc kubenswrapper[4989]: I1006 10:26:31.949444 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e276c28a-7ef5-4a0d-ae40-11603f6a647e" path="/var/lib/kubelet/pods/e276c28a-7ef5-4a0d-ae40-11603f6a647e/volumes" Oct 06 10:26:32 crc kubenswrapper[4989]: I1006 10:26:32.915175 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 10:26:33 crc kubenswrapper[4989]: I1006 10:26:33.617330 4989 generic.go:334] "Generic (PLEG): container finished" podID="b2be20a5-53e0-4fc6-a4d0-29206fbf5676" containerID="51e8997cfb34cbc8614a653d46de4ece77c1dd218cdaf51c108214506a3c35c8" exitCode=0 Oct 06 10:26:33 crc kubenswrapper[4989]: I1006 10:26:33.617448 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7tsmz" event={"ID":"b2be20a5-53e0-4fc6-a4d0-29206fbf5676","Type":"ContainerDied","Data":"51e8997cfb34cbc8614a653d46de4ece77c1dd218cdaf51c108214506a3c35c8"} Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.296811 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.394803 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data\") pod \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.394908 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gfhb\" (UniqueName: \"kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb\") pod \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.394942 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data\") pod \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.395017 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle\") pod \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\" (UID: \"b2be20a5-53e0-4fc6-a4d0-29206fbf5676\") " Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.402463 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "b2be20a5-53e0-4fc6-a4d0-29206fbf5676" (UID: "b2be20a5-53e0-4fc6-a4d0-29206fbf5676"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.404181 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb" (OuterVolumeSpecName: "kube-api-access-4gfhb") pod "b2be20a5-53e0-4fc6-a4d0-29206fbf5676" (UID: "b2be20a5-53e0-4fc6-a4d0-29206fbf5676"). InnerVolumeSpecName "kube-api-access-4gfhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.407624 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data" (OuterVolumeSpecName: "config-data") pod "b2be20a5-53e0-4fc6-a4d0-29206fbf5676" (UID: "b2be20a5-53e0-4fc6-a4d0-29206fbf5676"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.428080 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2be20a5-53e0-4fc6-a4d0-29206fbf5676" (UID: "b2be20a5-53e0-4fc6-a4d0-29206fbf5676"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.497955 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.498002 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gfhb\" (UniqueName: \"kubernetes.io/projected/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-kube-api-access-4gfhb\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.498013 4989 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.498021 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2be20a5-53e0-4fc6-a4d0-29206fbf5676-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.643774 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-7tsmz" event={"ID":"b2be20a5-53e0-4fc6-a4d0-29206fbf5676","Type":"ContainerDied","Data":"d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98"} Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.643820 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6e240a6acf6abaf355336d1428d72308766de033a75465b4b6ffcd93d23ab98" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.643899 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-7tsmz" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.877343 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 06 10:26:35 crc kubenswrapper[4989]: E1006 10:26:35.877837 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2be20a5-53e0-4fc6-a4d0-29206fbf5676" containerName="manila-db-sync" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.877860 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2be20a5-53e0-4fc6-a4d0-29206fbf5676" containerName="manila-db-sync" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.878178 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2be20a5-53e0-4fc6-a4d0-29206fbf5676" containerName="manila-db-sync" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.879478 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.881097 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.881164 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.884671 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.884859 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-2jwdh" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.911852 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.985136 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.986978 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.993251 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 06 10:26:35 crc kubenswrapper[4989]: I1006 10:26:35.997366 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.009844 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.009876 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.009912 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9jg\" (UniqueName: \"kubernetes.io/projected/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-kube-api-access-4x9jg\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.009956 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.010023 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-scripts\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.010087 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.073534 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.075213 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.105606 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112006 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112047 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112087 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112121 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-scripts\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112163 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112183 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112212 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9jg\" (UniqueName: \"kubernetes.io/projected/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-kube-api-access-4x9jg\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112238 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jvs9\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-kube-api-access-7jvs9\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112290 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-ceph\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112352 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112374 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112406 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-scripts\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112444 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.112555 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.121665 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.123630 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.125085 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-scripts\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.125634 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.134396 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9jg\" (UniqueName: \"kubernetes.io/projected/53c8f6a6-d2c2-4e4f-abed-30eb20208a7a-kube-api-access-4x9jg\") pod \"manila-scheduler-0\" (UID: \"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a\") " pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.200964 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.214969 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215034 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-scripts\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215062 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxnlb\" (UniqueName: \"kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215106 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215139 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jvs9\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-kube-api-access-7jvs9\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215185 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215567 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-ceph\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215606 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215664 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215729 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215890 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215953 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.215979 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.216829 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.216959 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/12cdf345-49c9-4a58-9357-7cdf2b7622eb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.218915 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-scripts\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.219883 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-ceph\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.227204 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.229836 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.232886 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.237242 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.242999 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.247390 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.248619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12cdf345-49c9-4a58-9357-7cdf2b7622eb-config-data\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.255186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jvs9\" (UniqueName: \"kubernetes.io/projected/12cdf345-49c9-4a58-9357-7cdf2b7622eb-kube-api-access-7jvs9\") pod \"manila-share-share1-0\" (UID: \"12cdf345-49c9-4a58-9357-7cdf2b7622eb\") " pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318171 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318252 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318294 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318326 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxnlb\" (UniqueName: \"kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318363 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318403 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4280fd-c098-4b4b-a926-123012f6b048-etc-machine-id\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318437 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkh6l\" (UniqueName: \"kubernetes.io/projected/1d4280fd-c098-4b4b-a926-123012f6b048-kube-api-access-zkh6l\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318500 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data-custom\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318531 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318567 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4280fd-c098-4b4b-a926-123012f6b048-logs\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.318606 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-scripts\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.319580 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.321492 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.321673 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.321775 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.333460 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.335860 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxnlb\" (UniqueName: \"kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb\") pod \"dnsmasq-dns-664779f8b9-xxvtr\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.399199 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420473 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420538 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420620 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4280fd-c098-4b4b-a926-123012f6b048-etc-machine-id\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420682 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkh6l\" (UniqueName: \"kubernetes.io/projected/1d4280fd-c098-4b4b-a926-123012f6b048-kube-api-access-zkh6l\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420700 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data-custom\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420735 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4280fd-c098-4b4b-a926-123012f6b048-logs\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420740 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1d4280fd-c098-4b4b-a926-123012f6b048-etc-machine-id\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.420762 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-scripts\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.422017 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d4280fd-c098-4b4b-a926-123012f6b048-logs\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.424456 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.425227 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-scripts\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.425802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.427307 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d4280fd-c098-4b4b-a926-123012f6b048-config-data-custom\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.441589 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkh6l\" (UniqueName: \"kubernetes.io/projected/1d4280fd-c098-4b4b-a926-123012f6b048-kube-api-access-zkh6l\") pod \"manila-api-0\" (UID: \"1d4280fd-c098-4b4b-a926-123012f6b048\") " pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.736188 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 06 10:26:36 crc kubenswrapper[4989]: I1006 10:26:36.813549 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 06 10:26:37 crc kubenswrapper[4989]: W1006 10:26:37.096000 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12cdf345_49c9_4a58_9357_7cdf2b7622eb.slice/crio-8772fbc4b9742ffb6e0bcd7b355c32b1d3ed364510d61b0749e5f265a997a23e WatchSource:0}: Error finding container 8772fbc4b9742ffb6e0bcd7b355c32b1d3ed364510d61b0749e5f265a997a23e: Status 404 returned error can't find the container with id 8772fbc4b9742ffb6e0bcd7b355c32b1d3ed364510d61b0749e5f265a997a23e Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.100151 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 06 10:26:37 crc kubenswrapper[4989]: W1006 10:26:37.105981 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod086d4d55_553a_4ee9_9de8_359850fc0256.slice/crio-673e0f013184072e540179cb842a08162af7cb342e79e53c839de0d3392fd0ec WatchSource:0}: Error finding container 673e0f013184072e540179cb842a08162af7cb342e79e53c839de0d3392fd0ec: Status 404 returned error can't find the container with id 673e0f013184072e540179cb842a08162af7cb342e79e53c839de0d3392fd0ec Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.110904 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.426945 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 06 10:26:37 crc kubenswrapper[4989]: W1006 10:26:37.467855 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d4280fd_c098_4b4b_a926_123012f6b048.slice/crio-a98d9366a848e1ac3f0e874fe34292480cff52e0db0f01c82611d46c7fdf8c9c WatchSource:0}: Error finding container a98d9366a848e1ac3f0e874fe34292480cff52e0db0f01c82611d46c7fdf8c9c: Status 404 returned error can't find the container with id a98d9366a848e1ac3f0e874fe34292480cff52e0db0f01c82611d46c7fdf8c9c Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.670072 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"12cdf345-49c9-4a58-9357-7cdf2b7622eb","Type":"ContainerStarted","Data":"8772fbc4b9742ffb6e0bcd7b355c32b1d3ed364510d61b0749e5f265a997a23e"} Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.674318 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1d4280fd-c098-4b4b-a926-123012f6b048","Type":"ContainerStarted","Data":"a98d9366a848e1ac3f0e874fe34292480cff52e0db0f01c82611d46c7fdf8c9c"} Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.678728 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a","Type":"ContainerStarted","Data":"188fcafa09e3813c40a22d24749e761590892d83ccb9cc6e4112ac4a030442f9"} Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.680698 4989 generic.go:334] "Generic (PLEG): container finished" podID="086d4d55-553a-4ee9-9de8-359850fc0256" containerID="1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f" exitCode=0 Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.680722 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" event={"ID":"086d4d55-553a-4ee9-9de8-359850fc0256","Type":"ContainerDied","Data":"1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f"} Oct 06 10:26:37 crc kubenswrapper[4989]: I1006 10:26:37.680747 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" event={"ID":"086d4d55-553a-4ee9-9de8-359850fc0256","Type":"ContainerStarted","Data":"673e0f013184072e540179cb842a08162af7cb342e79e53c839de0d3392fd0ec"} Oct 06 10:26:38 crc kubenswrapper[4989]: I1006 10:26:38.695159 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" event={"ID":"086d4d55-553a-4ee9-9de8-359850fc0256","Type":"ContainerStarted","Data":"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615"} Oct 06 10:26:38 crc kubenswrapper[4989]: I1006 10:26:38.695630 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:38 crc kubenswrapper[4989]: I1006 10:26:38.698941 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1d4280fd-c098-4b4b-a926-123012f6b048","Type":"ContainerStarted","Data":"473b2f4270f6aac30624791ccc933143f9cc6cb4dc159c95641a914c7ba6a634"} Oct 06 10:26:38 crc kubenswrapper[4989]: I1006 10:26:38.708222 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a","Type":"ContainerStarted","Data":"719ab48f57195ec4bc3bd8e3766070d61f683dd589b2f57d23cde83e595988b7"} Oct 06 10:26:38 crc kubenswrapper[4989]: I1006 10:26:38.717616 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" podStartSLOduration=2.717600377 podStartE2EDuration="2.717600377s" podCreationTimestamp="2025-10-06 10:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:26:38.714705413 +0000 UTC m=+6449.504730983" watchObservedRunningTime="2025-10-06 10:26:38.717600377 +0000 UTC m=+6449.507625957" Oct 06 10:26:39 crc kubenswrapper[4989]: I1006 10:26:39.721273 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"53c8f6a6-d2c2-4e4f-abed-30eb20208a7a","Type":"ContainerStarted","Data":"ddf5de3d739ac689201f0000ba124e1469cdf6e1a8503a7c7bda5819adb91396"} Oct 06 10:26:39 crc kubenswrapper[4989]: I1006 10:26:39.732796 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1d4280fd-c098-4b4b-a926-123012f6b048","Type":"ContainerStarted","Data":"3275fc84b78b87d3240f2f5a27ee54806945e6f592bd8f3c054ebe164ca89ffc"} Oct 06 10:26:39 crc kubenswrapper[4989]: I1006 10:26:39.732839 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 06 10:26:39 crc kubenswrapper[4989]: I1006 10:26:39.751213 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.080943881 podStartE2EDuration="4.751192192s" podCreationTimestamp="2025-10-06 10:26:35 +0000 UTC" firstStartedPulling="2025-10-06 10:26:36.84236948 +0000 UTC m=+6447.632395060" lastFinishedPulling="2025-10-06 10:26:37.512617791 +0000 UTC m=+6448.302643371" observedRunningTime="2025-10-06 10:26:39.736104597 +0000 UTC m=+6450.526130177" watchObservedRunningTime="2025-10-06 10:26:39.751192192 +0000 UTC m=+6450.541217772" Oct 06 10:26:39 crc kubenswrapper[4989]: I1006 10:26:39.778355 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.778310784 podStartE2EDuration="3.778310784s" podCreationTimestamp="2025-10-06 10:26:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:26:39.757968437 +0000 UTC m=+6450.547994017" watchObservedRunningTime="2025-10-06 10:26:39.778310784 +0000 UTC m=+6450.568336384" Oct 06 10:26:45 crc kubenswrapper[4989]: I1006 10:26:45.820310 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"12cdf345-49c9-4a58-9357-7cdf2b7622eb","Type":"ContainerStarted","Data":"1fbed05d00259d88b08ad96c629c154dce3b23dc381dfd7f0b780760958123bb"} Oct 06 10:26:45 crc kubenswrapper[4989]: I1006 10:26:45.820991 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"12cdf345-49c9-4a58-9357-7cdf2b7622eb","Type":"ContainerStarted","Data":"d02ed21f9bf0534370d400d3aaf26a79d9cbd4d4418783ced2be5b5f8cdd7c87"} Oct 06 10:26:45 crc kubenswrapper[4989]: I1006 10:26:45.845111 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.192227442 podStartE2EDuration="10.845096679s" podCreationTimestamp="2025-10-06 10:26:35 +0000 UTC" firstStartedPulling="2025-10-06 10:26:37.098207935 +0000 UTC m=+6447.888233515" lastFinishedPulling="2025-10-06 10:26:44.751077132 +0000 UTC m=+6455.541102752" observedRunningTime="2025-10-06 10:26:45.842026701 +0000 UTC m=+6456.632052291" watchObservedRunningTime="2025-10-06 10:26:45.845096679 +0000 UTC m=+6456.635122259" Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.204486 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.334007 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.401235 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.475258 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.475509 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="dnsmasq-dns" containerID="cri-o://f4cabf623c76ab6568f152e7040eac6c77fdc1d74f352e719e51039b404d1d2b" gracePeriod=10 Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.835700 4989 generic.go:334] "Generic (PLEG): container finished" podID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerID="f4cabf623c76ab6568f152e7040eac6c77fdc1d74f352e719e51039b404d1d2b" exitCode=0 Oct 06 10:26:46 crc kubenswrapper[4989]: I1006 10:26:46.837482 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" event={"ID":"fb0ff807-5082-4bde-b6c7-f359964df63a","Type":"ContainerDied","Data":"f4cabf623c76ab6568f152e7040eac6c77fdc1d74f352e719e51039b404d1d2b"} Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.078087 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.188409 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb\") pod \"fb0ff807-5082-4bde-b6c7-f359964df63a\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.188791 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config\") pod \"fb0ff807-5082-4bde-b6c7-f359964df63a\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.188841 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb\") pod \"fb0ff807-5082-4bde-b6c7-f359964df63a\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.188872 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2mwl\" (UniqueName: \"kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl\") pod \"fb0ff807-5082-4bde-b6c7-f359964df63a\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.188897 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc\") pod \"fb0ff807-5082-4bde-b6c7-f359964df63a\" (UID: \"fb0ff807-5082-4bde-b6c7-f359964df63a\") " Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.198864 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl" (OuterVolumeSpecName: "kube-api-access-x2mwl") pod "fb0ff807-5082-4bde-b6c7-f359964df63a" (UID: "fb0ff807-5082-4bde-b6c7-f359964df63a"). InnerVolumeSpecName "kube-api-access-x2mwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.258719 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config" (OuterVolumeSpecName: "config") pod "fb0ff807-5082-4bde-b6c7-f359964df63a" (UID: "fb0ff807-5082-4bde-b6c7-f359964df63a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.263462 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fb0ff807-5082-4bde-b6c7-f359964df63a" (UID: "fb0ff807-5082-4bde-b6c7-f359964df63a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.284615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb0ff807-5082-4bde-b6c7-f359964df63a" (UID: "fb0ff807-5082-4bde-b6c7-f359964df63a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.286200 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fb0ff807-5082-4bde-b6c7-f359964df63a" (UID: "fb0ff807-5082-4bde-b6c7-f359964df63a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.291564 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.291600 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.291612 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2mwl\" (UniqueName: \"kubernetes.io/projected/fb0ff807-5082-4bde-b6c7-f359964df63a-kube-api-access-x2mwl\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.291624 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.291645 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb0ff807-5082-4bde-b6c7-f359964df63a-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.847738 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.856394 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8cff6cc7-h6j8v" event={"ID":"fb0ff807-5082-4bde-b6c7-f359964df63a","Type":"ContainerDied","Data":"7605644d43c475bcedc324d0ea31f2c7235428a894d7f1e2301132c4d36e8a50"} Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.856613 4989 scope.go:117] "RemoveContainer" containerID="f4cabf623c76ab6568f152e7040eac6c77fdc1d74f352e719e51039b404d1d2b" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.929778 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.951645 4989 scope.go:117] "RemoveContainer" containerID="21fdedabdced0c757245207ded0d806336f2b5cdf44bdd87a691f5f7419c4637" Oct 06 10:26:47 crc kubenswrapper[4989]: I1006 10:26:47.972914 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c8cff6cc7-h6j8v"] Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.530902 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.532932 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="proxy-httpd" containerID="cri-o://550e55d51c1abe9cd0b737bd44613026589ccf004d3fc2c5ca975e3d397de4b8" gracePeriod=30 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.532914 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-central-agent" containerID="cri-o://e9478c871c99b639490e7dc6017f55105374bc3f4e965f914cc089a418dd3cf9" gracePeriod=30 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.533035 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="sg-core" containerID="cri-o://48911551e646ebe3307d957ccdd3f287fdf2aa3db598a31e2c52563cf464a005" gracePeriod=30 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.533048 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-notification-agent" containerID="cri-o://b4d5a7a243e105debe37a93dda262c322d01317642df4fd135ae924082b79b0f" gracePeriod=30 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.869194 4989 generic.go:334] "Generic (PLEG): container finished" podID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerID="550e55d51c1abe9cd0b737bd44613026589ccf004d3fc2c5ca975e3d397de4b8" exitCode=0 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.869224 4989 generic.go:334] "Generic (PLEG): container finished" podID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerID="48911551e646ebe3307d957ccdd3f287fdf2aa3db598a31e2c52563cf464a005" exitCode=2 Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.869243 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerDied","Data":"550e55d51c1abe9cd0b737bd44613026589ccf004d3fc2c5ca975e3d397de4b8"} Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.869270 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerDied","Data":"48911551e646ebe3307d957ccdd3f287fdf2aa3db598a31e2c52563cf464a005"} Oct 06 10:26:49 crc kubenswrapper[4989]: I1006 10:26:49.960423 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" path="/var/lib/kubelet/pods/fb0ff807-5082-4bde-b6c7-f359964df63a/volumes" Oct 06 10:26:50 crc kubenswrapper[4989]: I1006 10:26:50.890721 4989 generic.go:334] "Generic (PLEG): container finished" podID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerID="e9478c871c99b639490e7dc6017f55105374bc3f4e965f914cc089a418dd3cf9" exitCode=0 Oct 06 10:26:50 crc kubenswrapper[4989]: I1006 10:26:50.890824 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerDied","Data":"e9478c871c99b639490e7dc6017f55105374bc3f4e965f914cc089a418dd3cf9"} Oct 06 10:26:52 crc kubenswrapper[4989]: I1006 10:26:52.916056 4989 generic.go:334] "Generic (PLEG): container finished" podID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerID="b4d5a7a243e105debe37a93dda262c322d01317642df4fd135ae924082b79b0f" exitCode=0 Oct 06 10:26:52 crc kubenswrapper[4989]: I1006 10:26:52.916138 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerDied","Data":"b4d5a7a243e105debe37a93dda262c322d01317642df4fd135ae924082b79b0f"} Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.344847 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427463 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427504 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427558 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427605 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qxxf\" (UniqueName: \"kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427780 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427906 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.427931 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts\") pod \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\" (UID: \"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c\") " Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.429776 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.429815 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.436758 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts" (OuterVolumeSpecName: "scripts") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.438344 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf" (OuterVolumeSpecName: "kube-api-access-4qxxf") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "kube-api-access-4qxxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.465145 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.505845 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530126 4989 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530161 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530177 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qxxf\" (UniqueName: \"kubernetes.io/projected/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-kube-api-access-4qxxf\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530191 4989 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530200 4989 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.530211 4989 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-scripts\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.546416 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data" (OuterVolumeSpecName: "config-data") pod "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" (UID: "e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.632219 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.930235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c","Type":"ContainerDied","Data":"89c0d02f520206654e5c4058a2e47022e7eb4890f42565eda84b1905af0d1bbc"} Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.930292 4989 scope.go:117] "RemoveContainer" containerID="550e55d51c1abe9cd0b737bd44613026589ccf004d3fc2c5ca975e3d397de4b8" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.930461 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.978975 4989 scope.go:117] "RemoveContainer" containerID="48911551e646ebe3307d957ccdd3f287fdf2aa3db598a31e2c52563cf464a005" Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.983184 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:53 crc kubenswrapper[4989]: I1006 10:26:53.996484 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.008860 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009439 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="sg-core" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009456 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="sg-core" Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009474 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="init" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009481 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="init" Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009491 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="proxy-httpd" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009500 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="proxy-httpd" Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009522 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="dnsmasq-dns" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009529 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="dnsmasq-dns" Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009549 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-notification-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009556 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-notification-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: E1006 10:26:54.009621 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-central-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009630 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-central-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009880 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="sg-core" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009904 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb0ff807-5082-4bde-b6c7-f359964df63a" containerName="dnsmasq-dns" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009920 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-central-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009938 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="ceilometer-notification-agent" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.009949 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" containerName="proxy-httpd" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.012430 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.014067 4989 scope.go:117] "RemoveContainer" containerID="b4d5a7a243e105debe37a93dda262c322d01317642df4fd135ae924082b79b0f" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.017316 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.017478 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.046029 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.062152 4989 scope.go:117] "RemoveContainer" containerID="e9478c871c99b639490e7dc6017f55105374bc3f4e965f914cc089a418dd3cf9" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150279 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-log-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150382 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95st2\" (UniqueName: \"kubernetes.io/projected/d3514350-1ff6-43b2-b91b-a3781cb5050d-kube-api-access-95st2\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150551 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-run-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150578 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150596 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-config-data\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150611 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-scripts\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.150632 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252331 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-run-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252394 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252417 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-config-data\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252438 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-scripts\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252467 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252517 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-log-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252578 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95st2\" (UniqueName: \"kubernetes.io/projected/d3514350-1ff6-43b2-b91b-a3781cb5050d-kube-api-access-95st2\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.252944 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-run-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.253014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3514350-1ff6-43b2-b91b-a3781cb5050d-log-httpd\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.258101 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-config-data\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.258735 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.260636 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-scripts\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.278422 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3514350-1ff6-43b2-b91b-a3781cb5050d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.281318 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95st2\" (UniqueName: \"kubernetes.io/projected/d3514350-1ff6-43b2-b91b-a3781cb5050d-kube-api-access-95st2\") pod \"ceilometer-0\" (UID: \"d3514350-1ff6-43b2-b91b-a3781cb5050d\") " pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.353591 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.828828 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 06 10:26:54 crc kubenswrapper[4989]: I1006 10:26:54.942709 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3514350-1ff6-43b2-b91b-a3781cb5050d","Type":"ContainerStarted","Data":"d89b76679d64eb888c3f9bb0a78861f68183d0e9fbdbac0a8e65e11212889386"} Oct 06 10:26:55 crc kubenswrapper[4989]: I1006 10:26:55.959910 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c" path="/var/lib/kubelet/pods/e3d142f3-f0c3-4c8d-a29e-2c2be7834b5c/volumes" Oct 06 10:26:55 crc kubenswrapper[4989]: I1006 10:26:55.970332 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3514350-1ff6-43b2-b91b-a3781cb5050d","Type":"ContainerStarted","Data":"b9019bd8c89d6847f6e98ccb3801fd0d1034a803e135e8e97631c187504bdfe5"} Oct 06 10:26:56 crc kubenswrapper[4989]: I1006 10:26:56.980235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3514350-1ff6-43b2-b91b-a3781cb5050d","Type":"ContainerStarted","Data":"63c30cda78aec31d881ca294b065531446e3162fb41f62d00ddd135fd3321701"} Oct 06 10:26:57 crc kubenswrapper[4989]: I1006 10:26:57.816749 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 06 10:26:57 crc kubenswrapper[4989]: I1006 10:26:57.994082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3514350-1ff6-43b2-b91b-a3781cb5050d","Type":"ContainerStarted","Data":"934c3205b0af38066d099ea3b49a68b711da688a6d96de7f6ba81e560c5f0355"} Oct 06 10:26:58 crc kubenswrapper[4989]: I1006 10:26:58.121772 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 06 10:26:58 crc kubenswrapper[4989]: I1006 10:26:58.318580 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 06 10:26:59 crc kubenswrapper[4989]: I1006 10:26:59.007691 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3514350-1ff6-43b2-b91b-a3781cb5050d","Type":"ContainerStarted","Data":"8b9e221a04d63754cb37bab6eb460c291da5bbb46b2604c36d86799b0152d309"} Oct 06 10:26:59 crc kubenswrapper[4989]: I1006 10:26:59.009590 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 06 10:26:59 crc kubenswrapper[4989]: I1006 10:26:59.027885 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.272266284 podStartE2EDuration="6.027869176s" podCreationTimestamp="2025-10-06 10:26:53 +0000 UTC" firstStartedPulling="2025-10-06 10:26:54.844456612 +0000 UTC m=+6465.634482192" lastFinishedPulling="2025-10-06 10:26:58.600059504 +0000 UTC m=+6469.390085084" observedRunningTime="2025-10-06 10:26:59.025460806 +0000 UTC m=+6469.815486386" watchObservedRunningTime="2025-10-06 10:26:59.027869176 +0000 UTC m=+6469.817894756" Oct 06 10:27:03 crc kubenswrapper[4989]: I1006 10:27:03.935836 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:27:03 crc kubenswrapper[4989]: I1006 10:27:03.936284 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:27:24 crc kubenswrapper[4989]: I1006 10:27:24.326962 4989 scope.go:117] "RemoveContainer" containerID="ef12b1feead78763aac37daf3045e69df1156a52ac43b34dad9df053685b0003" Oct 06 10:27:24 crc kubenswrapper[4989]: I1006 10:27:24.367642 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 06 10:27:24 crc kubenswrapper[4989]: I1006 10:27:24.437929 4989 scope.go:117] "RemoveContainer" containerID="6191a08b492a84f18b8fc40aba656a9636d305478e190fa06977403750679adc" Oct 06 10:27:33 crc kubenswrapper[4989]: I1006 10:27:33.935156 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:27:33 crc kubenswrapper[4989]: I1006 10:27:33.935853 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.290922 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.300199 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.303750 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.316812 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427083 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427480 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fqhk\" (UniqueName: \"kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427504 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427592 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.427623 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.529549 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.529791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fqhk\" (UniqueName: \"kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.529827 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.529883 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.530085 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.530472 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.530909 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.530959 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.530960 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.531022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.531680 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.551458 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fqhk\" (UniqueName: \"kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk\") pod \"dnsmasq-dns-d45b67487-9s7j6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:43 crc kubenswrapper[4989]: I1006 10:27:43.635495 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:44 crc kubenswrapper[4989]: I1006 10:27:44.140364 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:27:44 crc kubenswrapper[4989]: I1006 10:27:44.524598 4989 generic.go:334] "Generic (PLEG): container finished" podID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerID="e8184ddfea371c61b67bfa09e27b292f46465403253464db77e0095c5b1d51ab" exitCode=0 Oct 06 10:27:44 crc kubenswrapper[4989]: I1006 10:27:44.524813 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" event={"ID":"63cd77cb-95d6-49dd-bb77-7e533c3743b6","Type":"ContainerDied","Data":"e8184ddfea371c61b67bfa09e27b292f46465403253464db77e0095c5b1d51ab"} Oct 06 10:27:44 crc kubenswrapper[4989]: I1006 10:27:44.525070 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" event={"ID":"63cd77cb-95d6-49dd-bb77-7e533c3743b6","Type":"ContainerStarted","Data":"40d3fde221ad7e14c9031e0591f0d0b57e06507e30542b1f1039edee4b6ca47c"} Oct 06 10:27:45 crc kubenswrapper[4989]: I1006 10:27:45.540232 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" event={"ID":"63cd77cb-95d6-49dd-bb77-7e533c3743b6","Type":"ContainerStarted","Data":"a848a248aee5208ff56f0799d9d371715037446c22e3c033b752fae9d83e2ef9"} Oct 06 10:27:45 crc kubenswrapper[4989]: I1006 10:27:45.541237 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:45 crc kubenswrapper[4989]: I1006 10:27:45.571097 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" podStartSLOduration=2.571073964 podStartE2EDuration="2.571073964s" podCreationTimestamp="2025-10-06 10:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:27:45.563573378 +0000 UTC m=+6516.353598978" watchObservedRunningTime="2025-10-06 10:27:45.571073964 +0000 UTC m=+6516.361099564" Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.636883 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.727887 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.728328 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="dnsmasq-dns" containerID="cri-o://6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615" gracePeriod=10 Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.955015 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cdb957947-rbss9"] Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.957726 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:53 crc kubenswrapper[4989]: I1006 10:27:53.967516 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cdb957947-rbss9"] Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.085471 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-openstack-cell1\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.085780 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-config\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.085870 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mksfk\" (UniqueName: \"kubernetes.io/projected/b35e50fb-921f-4cbe-adc6-6205aa5150f4-kube-api-access-mksfk\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.085890 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-nb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.085915 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.086015 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-dns-svc\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.187762 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-dns-svc\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.187887 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-openstack-cell1\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.187925 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-config\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.188007 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mksfk\" (UniqueName: \"kubernetes.io/projected/b35e50fb-921f-4cbe-adc6-6205aa5150f4-kube-api-access-mksfk\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.188037 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-nb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.188065 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.189357 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-config\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.190224 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-dns-svc\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.190733 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-openstack-cell1\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.193049 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-nb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.193169 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b35e50fb-921f-4cbe-adc6-6205aa5150f4-ovsdbserver-sb\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.217924 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mksfk\" (UniqueName: \"kubernetes.io/projected/b35e50fb-921f-4cbe-adc6-6205aa5150f4-kube-api-access-mksfk\") pod \"dnsmasq-dns-7cdb957947-rbss9\" (UID: \"b35e50fb-921f-4cbe-adc6-6205aa5150f4\") " pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.284065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.438315 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.495210 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxnlb\" (UniqueName: \"kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb\") pod \"086d4d55-553a-4ee9-9de8-359850fc0256\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.495553 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb\") pod \"086d4d55-553a-4ee9-9de8-359850fc0256\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.495596 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config\") pod \"086d4d55-553a-4ee9-9de8-359850fc0256\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.495696 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb\") pod \"086d4d55-553a-4ee9-9de8-359850fc0256\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.495725 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc\") pod \"086d4d55-553a-4ee9-9de8-359850fc0256\" (UID: \"086d4d55-553a-4ee9-9de8-359850fc0256\") " Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.504098 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb" (OuterVolumeSpecName: "kube-api-access-xxnlb") pod "086d4d55-553a-4ee9-9de8-359850fc0256" (UID: "086d4d55-553a-4ee9-9de8-359850fc0256"). InnerVolumeSpecName "kube-api-access-xxnlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.580401 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "086d4d55-553a-4ee9-9de8-359850fc0256" (UID: "086d4d55-553a-4ee9-9de8-359850fc0256"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.583312 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "086d4d55-553a-4ee9-9de8-359850fc0256" (UID: "086d4d55-553a-4ee9-9de8-359850fc0256"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.589860 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config" (OuterVolumeSpecName: "config") pod "086d4d55-553a-4ee9-9de8-359850fc0256" (UID: "086d4d55-553a-4ee9-9de8-359850fc0256"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.598816 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.598855 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.598868 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxnlb\" (UniqueName: \"kubernetes.io/projected/086d4d55-553a-4ee9-9de8-359850fc0256-kube-api-access-xxnlb\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.598882 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.608868 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "086d4d55-553a-4ee9-9de8-359850fc0256" (UID: "086d4d55-553a-4ee9-9de8-359850fc0256"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.647408 4989 generic.go:334] "Generic (PLEG): container finished" podID="086d4d55-553a-4ee9-9de8-359850fc0256" containerID="6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615" exitCode=0 Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.647446 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" event={"ID":"086d4d55-553a-4ee9-9de8-359850fc0256","Type":"ContainerDied","Data":"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615"} Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.647472 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" event={"ID":"086d4d55-553a-4ee9-9de8-359850fc0256","Type":"ContainerDied","Data":"673e0f013184072e540179cb842a08162af7cb342e79e53c839de0d3392fd0ec"} Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.647482 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-664779f8b9-xxvtr" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.647507 4989 scope.go:117] "RemoveContainer" containerID="6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.672333 4989 scope.go:117] "RemoveContainer" containerID="1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.692932 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.696289 4989 scope.go:117] "RemoveContainer" containerID="6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615" Oct 06 10:27:54 crc kubenswrapper[4989]: E1006 10:27:54.698055 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615\": container with ID starting with 6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615 not found: ID does not exist" containerID="6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.698092 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615"} err="failed to get container status \"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615\": rpc error: code = NotFound desc = could not find container \"6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615\": container with ID starting with 6d467dcb7d7a6c9b3e9623e90445ea3b4ec58ff04de2393062175f7a908ea615 not found: ID does not exist" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.698121 4989 scope.go:117] "RemoveContainer" containerID="1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f" Oct 06 10:27:54 crc kubenswrapper[4989]: E1006 10:27:54.698468 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f\": container with ID starting with 1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f not found: ID does not exist" containerID="1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.698500 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f"} err="failed to get container status \"1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f\": rpc error: code = NotFound desc = could not find container \"1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f\": container with ID starting with 1aa68d8319516a8df27845353ad95ccb492585f8442455ed74aa94ef7568310f not found: ID does not exist" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.700352 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086d4d55-553a-4ee9-9de8-359850fc0256-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.703748 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-664779f8b9-xxvtr"] Oct 06 10:27:54 crc kubenswrapper[4989]: I1006 10:27:54.812942 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cdb957947-rbss9"] Oct 06 10:27:55 crc kubenswrapper[4989]: I1006 10:27:55.663801 4989 generic.go:334] "Generic (PLEG): container finished" podID="b35e50fb-921f-4cbe-adc6-6205aa5150f4" containerID="f6dd4c13089c19ffea2c8fdaff13a438342171a9b1c4d2a3d98c3f4cf7d170f6" exitCode=0 Oct 06 10:27:55 crc kubenswrapper[4989]: I1006 10:27:55.664116 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" event={"ID":"b35e50fb-921f-4cbe-adc6-6205aa5150f4","Type":"ContainerDied","Data":"f6dd4c13089c19ffea2c8fdaff13a438342171a9b1c4d2a3d98c3f4cf7d170f6"} Oct 06 10:27:55 crc kubenswrapper[4989]: I1006 10:27:55.664137 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" event={"ID":"b35e50fb-921f-4cbe-adc6-6205aa5150f4","Type":"ContainerStarted","Data":"a4d7c70034249653ed5b269f9c05ad0cf8aca15b170e5b856ccea42a7598c1f0"} Oct 06 10:27:55 crc kubenswrapper[4989]: I1006 10:27:55.952280 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" path="/var/lib/kubelet/pods/086d4d55-553a-4ee9-9de8-359850fc0256/volumes" Oct 06 10:27:56 crc kubenswrapper[4989]: I1006 10:27:56.674184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" event={"ID":"b35e50fb-921f-4cbe-adc6-6205aa5150f4","Type":"ContainerStarted","Data":"d39ec0950c2f592250f6f688b2119fc9966dfb2e863096109ad1aca37b1e8a2a"} Oct 06 10:27:56 crc kubenswrapper[4989]: I1006 10:27:56.674353 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:27:56 crc kubenswrapper[4989]: I1006 10:27:56.700016 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" podStartSLOduration=3.699997445 podStartE2EDuration="3.699997445s" podCreationTimestamp="2025-10-06 10:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:27:56.690514591 +0000 UTC m=+6527.480540171" watchObservedRunningTime="2025-10-06 10:27:56.699997445 +0000 UTC m=+6527.490023035" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.048451 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd"] Oct 06 10:28:00 crc kubenswrapper[4989]: E1006 10:28:00.049555 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="init" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.049576 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="init" Oct 06 10:28:00 crc kubenswrapper[4989]: E1006 10:28:00.049625 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="dnsmasq-dns" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.049634 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="dnsmasq-dns" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.049922 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="086d4d55-553a-4ee9-9de8-359850fc0256" containerName="dnsmasq-dns" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.051132 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.053227 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.053805 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.053969 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.055707 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.080332 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd"] Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.140126 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.140178 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.140410 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.140629 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kpht\" (UniqueName: \"kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.141060 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.243275 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.243369 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.243405 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.243482 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.243546 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kpht\" (UniqueName: \"kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.252322 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.252843 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.254703 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.258765 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.269051 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kpht\" (UniqueName: \"kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.380704 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:00 crc kubenswrapper[4989]: I1006 10:28:00.938715 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd"] Oct 06 10:28:01 crc kubenswrapper[4989]: I1006 10:28:01.729003 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" event={"ID":"a78e83ee-accd-45a6-be14-5d73b7ce2060","Type":"ContainerStarted","Data":"28ffd3b6088842c4c4f42ca2e384b2984d53e33ae5ec64e7e4a5751e3536dedf"} Oct 06 10:28:03 crc kubenswrapper[4989]: I1006 10:28:03.935341 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:28:03 crc kubenswrapper[4989]: I1006 10:28:03.935973 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:28:03 crc kubenswrapper[4989]: I1006 10:28:03.948844 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:28:03 crc kubenswrapper[4989]: I1006 10:28:03.949609 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:28:03 crc kubenswrapper[4989]: I1006 10:28:03.949692 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46" gracePeriod=600 Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.285825 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cdb957947-rbss9" Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.352485 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.354508 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="dnsmasq-dns" containerID="cri-o://a848a248aee5208ff56f0799d9d371715037446c22e3c033b752fae9d83e2ef9" gracePeriod=10 Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.774572 4989 generic.go:334] "Generic (PLEG): container finished" podID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerID="a848a248aee5208ff56f0799d9d371715037446c22e3c033b752fae9d83e2ef9" exitCode=0 Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.775108 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" event={"ID":"63cd77cb-95d6-49dd-bb77-7e533c3743b6","Type":"ContainerDied","Data":"a848a248aee5208ff56f0799d9d371715037446c22e3c033b752fae9d83e2ef9"} Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.778684 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46" exitCode=0 Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.778724 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46"} Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.778745 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d"} Oct 06 10:28:04 crc kubenswrapper[4989]: I1006 10:28:04.778801 4989 scope.go:117] "RemoveContainer" containerID="9eb524421c8c109cdf9a685d734b64a9890dcbd1c1eec02ace55a90632fdbadb" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.019507 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.155937 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.156384 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.156478 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fqhk\" (UniqueName: \"kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.156620 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.156702 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.156809 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb\") pod \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\" (UID: \"63cd77cb-95d6-49dd-bb77-7e533c3743b6\") " Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.184568 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk" (OuterVolumeSpecName: "kube-api-access-8fqhk") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "kube-api-access-8fqhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.226195 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.227301 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.230014 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.241312 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.248188 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config" (OuterVolumeSpecName: "config") pod "63cd77cb-95d6-49dd-bb77-7e533c3743b6" (UID: "63cd77cb-95d6-49dd-bb77-7e533c3743b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259105 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fqhk\" (UniqueName: \"kubernetes.io/projected/63cd77cb-95d6-49dd-bb77-7e533c3743b6-kube-api-access-8fqhk\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259144 4989 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259158 4989 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259169 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259184 4989 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.259195 4989 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63cd77cb-95d6-49dd-bb77-7e533c3743b6-config\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.797721 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" event={"ID":"63cd77cb-95d6-49dd-bb77-7e533c3743b6","Type":"ContainerDied","Data":"40d3fde221ad7e14c9031e0591f0d0b57e06507e30542b1f1039edee4b6ca47c"} Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.797790 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d45b67487-9s7j6" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.798043 4989 scope.go:117] "RemoveContainer" containerID="a848a248aee5208ff56f0799d9d371715037446c22e3c033b752fae9d83e2ef9" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.825271 4989 scope.go:117] "RemoveContainer" containerID="e8184ddfea371c61b67bfa09e27b292f46465403253464db77e0095c5b1d51ab" Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.845284 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.856329 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d45b67487-9s7j6"] Oct 06 10:28:05 crc kubenswrapper[4989]: I1006 10:28:05.979882 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" path="/var/lib/kubelet/pods/63cd77cb-95d6-49dd-bb77-7e533c3743b6/volumes" Oct 06 10:28:13 crc kubenswrapper[4989]: I1006 10:28:13.202898 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:28:13 crc kubenswrapper[4989]: I1006 10:28:13.878966 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" event={"ID":"a78e83ee-accd-45a6-be14-5d73b7ce2060","Type":"ContainerStarted","Data":"900aef3917e98b76b2660abceedcc95e1f75d0d49f23763a0759416a9c69eb6c"} Oct 06 10:28:13 crc kubenswrapper[4989]: I1006 10:28:13.908178 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" podStartSLOduration=1.654376424 podStartE2EDuration="13.908157011s" podCreationTimestamp="2025-10-06 10:28:00 +0000 UTC" firstStartedPulling="2025-10-06 10:28:00.946546149 +0000 UTC m=+6531.736571749" lastFinishedPulling="2025-10-06 10:28:13.200326746 +0000 UTC m=+6543.990352336" observedRunningTime="2025-10-06 10:28:13.904813584 +0000 UTC m=+6544.694839164" watchObservedRunningTime="2025-10-06 10:28:13.908157011 +0000 UTC m=+6544.698182591" Oct 06 10:28:24 crc kubenswrapper[4989]: I1006 10:28:24.559572 4989 scope.go:117] "RemoveContainer" containerID="920efe9d85bc10aecfa0897062089e0e9b85f7bdbb880e2d66ddb48aec0937ba" Oct 06 10:28:24 crc kubenswrapper[4989]: I1006 10:28:24.936839 4989 scope.go:117] "RemoveContainer" containerID="7ca8944d8a3552d46c08819280437e46fb83766e451cda7aa935c60704cdf59c" Oct 06 10:28:24 crc kubenswrapper[4989]: I1006 10:28:24.964108 4989 scope.go:117] "RemoveContainer" containerID="f8c909ea9a4843ca480af0ad547dcb57df96bd5ebf3752bb7606f73681afcb86" Oct 06 10:28:24 crc kubenswrapper[4989]: I1006 10:28:24.990019 4989 scope.go:117] "RemoveContainer" containerID="1bc81d3a014ed55b5dfe4a0a40e3c7b54446aa020e598da857f402e485de772b" Oct 06 10:28:27 crc kubenswrapper[4989]: I1006 10:28:27.049791 4989 generic.go:334] "Generic (PLEG): container finished" podID="a78e83ee-accd-45a6-be14-5d73b7ce2060" containerID="900aef3917e98b76b2660abceedcc95e1f75d0d49f23763a0759416a9c69eb6c" exitCode=0 Oct 06 10:28:27 crc kubenswrapper[4989]: I1006 10:28:27.049847 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" event={"ID":"a78e83ee-accd-45a6-be14-5d73b7ce2060","Type":"ContainerDied","Data":"900aef3917e98b76b2660abceedcc95e1f75d0d49f23763a0759416a9c69eb6c"} Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.490312 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.525603 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key\") pod \"a78e83ee-accd-45a6-be14-5d73b7ce2060\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.525697 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory\") pod \"a78e83ee-accd-45a6-be14-5d73b7ce2060\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.525943 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle\") pod \"a78e83ee-accd-45a6-be14-5d73b7ce2060\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.526003 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kpht\" (UniqueName: \"kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht\") pod \"a78e83ee-accd-45a6-be14-5d73b7ce2060\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.526099 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph\") pod \"a78e83ee-accd-45a6-be14-5d73b7ce2060\" (UID: \"a78e83ee-accd-45a6-be14-5d73b7ce2060\") " Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.534877 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph" (OuterVolumeSpecName: "ceph") pod "a78e83ee-accd-45a6-be14-5d73b7ce2060" (UID: "a78e83ee-accd-45a6-be14-5d73b7ce2060"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.534921 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht" (OuterVolumeSpecName: "kube-api-access-2kpht") pod "a78e83ee-accd-45a6-be14-5d73b7ce2060" (UID: "a78e83ee-accd-45a6-be14-5d73b7ce2060"). InnerVolumeSpecName "kube-api-access-2kpht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.535992 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "a78e83ee-accd-45a6-be14-5d73b7ce2060" (UID: "a78e83ee-accd-45a6-be14-5d73b7ce2060"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.564242 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a78e83ee-accd-45a6-be14-5d73b7ce2060" (UID: "a78e83ee-accd-45a6-be14-5d73b7ce2060"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.586798 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory" (OuterVolumeSpecName: "inventory") pod "a78e83ee-accd-45a6-be14-5d73b7ce2060" (UID: "a78e83ee-accd-45a6-be14-5d73b7ce2060"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.628025 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.628354 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.628364 4989 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.628378 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kpht\" (UniqueName: \"kubernetes.io/projected/a78e83ee-accd-45a6-be14-5d73b7ce2060-kube-api-access-2kpht\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:28 crc kubenswrapper[4989]: I1006 10:28:28.628388 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a78e83ee-accd-45a6-be14-5d73b7ce2060-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:28:29 crc kubenswrapper[4989]: I1006 10:28:29.073919 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" event={"ID":"a78e83ee-accd-45a6-be14-5d73b7ce2060","Type":"ContainerDied","Data":"28ffd3b6088842c4c4f42ca2e384b2984d53e33ae5ec64e7e4a5751e3536dedf"} Oct 06 10:28:29 crc kubenswrapper[4989]: I1006 10:28:29.073968 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ffd3b6088842c4c4f42ca2e384b2984d53e33ae5ec64e7e4a5751e3536dedf" Oct 06 10:28:29 crc kubenswrapper[4989]: I1006 10:28:29.074005 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.518308 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw"] Oct 06 10:28:37 crc kubenswrapper[4989]: E1006 10:28:37.523764 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="dnsmasq-dns" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.524005 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="dnsmasq-dns" Oct 06 10:28:37 crc kubenswrapper[4989]: E1006 10:28:37.524100 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="init" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.524160 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="init" Oct 06 10:28:37 crc kubenswrapper[4989]: E1006 10:28:37.524246 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78e83ee-accd-45a6-be14-5d73b7ce2060" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.524298 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78e83ee-accd-45a6-be14-5d73b7ce2060" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.524589 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78e83ee-accd-45a6-be14-5d73b7ce2060" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.524688 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="63cd77cb-95d6-49dd-bb77-7e533c3743b6" containerName="dnsmasq-dns" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.525495 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.527674 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.528667 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.529008 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.531422 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.541374 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw"] Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.650058 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.650473 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.650636 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.650697 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gttrl\" (UniqueName: \"kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.650843 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.752909 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.752983 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.753056 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.753211 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.753260 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gttrl\" (UniqueName: \"kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.760635 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.760722 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.761240 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.768461 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.775966 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gttrl\" (UniqueName: \"kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:37 crc kubenswrapper[4989]: I1006 10:28:37.849161 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:28:38 crc kubenswrapper[4989]: I1006 10:28:38.474715 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw"] Oct 06 10:28:38 crc kubenswrapper[4989]: I1006 10:28:38.483971 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:28:39 crc kubenswrapper[4989]: I1006 10:28:39.205995 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" event={"ID":"5e33f7f9-1020-4783-bee8-c3effc5defb2","Type":"ContainerStarted","Data":"dad1ed2fb8f33d2d1617279207f601ce6728e884bb21616ef23fbf4aa733cda2"} Oct 06 10:28:39 crc kubenswrapper[4989]: I1006 10:28:39.206742 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" event={"ID":"5e33f7f9-1020-4783-bee8-c3effc5defb2","Type":"ContainerStarted","Data":"37f02cbf8490201e2be0586ae92faa0f6b4af24ffe4eaff4e5660eb7351c5580"} Oct 06 10:28:39 crc kubenswrapper[4989]: I1006 10:28:39.230764 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" podStartSLOduration=2.025565951 podStartE2EDuration="2.230738596s" podCreationTimestamp="2025-10-06 10:28:37 +0000 UTC" firstStartedPulling="2025-10-06 10:28:38.483794465 +0000 UTC m=+6569.273820045" lastFinishedPulling="2025-10-06 10:28:38.68896711 +0000 UTC m=+6569.478992690" observedRunningTime="2025-10-06 10:28:39.222252852 +0000 UTC m=+6570.012278442" watchObservedRunningTime="2025-10-06 10:28:39.230738596 +0000 UTC m=+6570.020764186" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.705820 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.724599 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.735007 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.867184 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.867541 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.867635 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wx6h\" (UniqueName: \"kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.970994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.971541 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.971733 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wx6h\" (UniqueName: \"kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.971802 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.972039 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:57 crc kubenswrapper[4989]: I1006 10:28:57.991445 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wx6h\" (UniqueName: \"kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h\") pod \"community-operators-l9k5q\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:58 crc kubenswrapper[4989]: I1006 10:28:58.060932 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:28:58 crc kubenswrapper[4989]: I1006 10:28:58.558830 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.106843 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.111549 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.114552 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.114630 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.114866 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8fj6\" (UniqueName: \"kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.128122 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.216777 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8fj6\" (UniqueName: \"kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.217540 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.217715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.218068 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.218231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.239678 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8fj6\" (UniqueName: \"kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6\") pod \"redhat-marketplace-drjhz\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.427355 4989 generic.go:334] "Generic (PLEG): container finished" podID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerID="60984051408d41785419b7e3647ba7fe66c5e69227d778b64e82a9e5d79e3c78" exitCode=0 Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.427555 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerDied","Data":"60984051408d41785419b7e3647ba7fe66c5e69227d778b64e82a9e5d79e3c78"} Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.427577 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerStarted","Data":"cb3faf0e8de663063b46c981946c5cc016c21fc2d38390e8fca0b1fd42542764"} Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.441421 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:28:59 crc kubenswrapper[4989]: W1006 10:28:59.925875 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d7b2894_777e_4bbe_a290_a27012f07476.slice/crio-530295e7454d2f21667c8fda38ed16989a1c3b65b519abb3c84bb2b32786169b WatchSource:0}: Error finding container 530295e7454d2f21667c8fda38ed16989a1c3b65b519abb3c84bb2b32786169b: Status 404 returned error can't find the container with id 530295e7454d2f21667c8fda38ed16989a1c3b65b519abb3c84bb2b32786169b Oct 06 10:28:59 crc kubenswrapper[4989]: I1006 10:28:59.927317 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:29:00 crc kubenswrapper[4989]: I1006 10:29:00.438525 4989 generic.go:334] "Generic (PLEG): container finished" podID="7d7b2894-777e-4bbe-a290-a27012f07476" containerID="a514d0159a6b6a3d7d1e810bb9abfbd20bc6c7f288137ec12e2027b768dbba10" exitCode=0 Oct 06 10:29:00 crc kubenswrapper[4989]: I1006 10:29:00.438608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerDied","Data":"a514d0159a6b6a3d7d1e810bb9abfbd20bc6c7f288137ec12e2027b768dbba10"} Oct 06 10:29:00 crc kubenswrapper[4989]: I1006 10:29:00.438883 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerStarted","Data":"530295e7454d2f21667c8fda38ed16989a1c3b65b519abb3c84bb2b32786169b"} Oct 06 10:29:01 crc kubenswrapper[4989]: I1006 10:29:01.460130 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerStarted","Data":"8ff660243f1660d251d6b26304586e1d0b2a54162f24006a60bec92cd9a9c6fa"} Oct 06 10:29:02 crc kubenswrapper[4989]: I1006 10:29:02.477934 4989 generic.go:334] "Generic (PLEG): container finished" podID="7d7b2894-777e-4bbe-a290-a27012f07476" containerID="2c79f94264f32d153dd7f4ec4bc3429f04f6bd6ea8e59d40d4b35dc5795d4b4a" exitCode=0 Oct 06 10:29:02 crc kubenswrapper[4989]: I1006 10:29:02.478057 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerDied","Data":"2c79f94264f32d153dd7f4ec4bc3429f04f6bd6ea8e59d40d4b35dc5795d4b4a"} Oct 06 10:29:02 crc kubenswrapper[4989]: I1006 10:29:02.481390 4989 generic.go:334] "Generic (PLEG): container finished" podID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerID="8ff660243f1660d251d6b26304586e1d0b2a54162f24006a60bec92cd9a9c6fa" exitCode=0 Oct 06 10:29:02 crc kubenswrapper[4989]: I1006 10:29:02.481432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerDied","Data":"8ff660243f1660d251d6b26304586e1d0b2a54162f24006a60bec92cd9a9c6fa"} Oct 06 10:29:03 crc kubenswrapper[4989]: I1006 10:29:03.496117 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerStarted","Data":"7c030879e5b59543b1c061f84727b2f73cba0187d1786487a64d1b3e7fcec891"} Oct 06 10:29:03 crc kubenswrapper[4989]: I1006 10:29:03.499971 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerStarted","Data":"04f9be063e8fa447d08b7674e22491d12dc35f5866e94186d584a34b72407c2f"} Oct 06 10:29:03 crc kubenswrapper[4989]: I1006 10:29:03.518088 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l9k5q" podStartSLOduration=2.9949824019999998 podStartE2EDuration="6.518069422s" podCreationTimestamp="2025-10-06 10:28:57 +0000 UTC" firstStartedPulling="2025-10-06 10:28:59.429379188 +0000 UTC m=+6590.219404758" lastFinishedPulling="2025-10-06 10:29:02.952466188 +0000 UTC m=+6593.742491778" observedRunningTime="2025-10-06 10:29:03.516101886 +0000 UTC m=+6594.306127496" watchObservedRunningTime="2025-10-06 10:29:03.518069422 +0000 UTC m=+6594.308095002" Oct 06 10:29:03 crc kubenswrapper[4989]: I1006 10:29:03.548571 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-drjhz" podStartSLOduration=1.853321327 podStartE2EDuration="4.548541111s" podCreationTimestamp="2025-10-06 10:28:59 +0000 UTC" firstStartedPulling="2025-10-06 10:29:00.440679751 +0000 UTC m=+6591.230705331" lastFinishedPulling="2025-10-06 10:29:03.135899525 +0000 UTC m=+6593.925925115" observedRunningTime="2025-10-06 10:29:03.540726885 +0000 UTC m=+6594.330752465" watchObservedRunningTime="2025-10-06 10:29:03.548541111 +0000 UTC m=+6594.338566701" Oct 06 10:29:08 crc kubenswrapper[4989]: I1006 10:29:08.062053 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:08 crc kubenswrapper[4989]: I1006 10:29:08.063549 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:08 crc kubenswrapper[4989]: I1006 10:29:08.107402 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:08 crc kubenswrapper[4989]: I1006 10:29:08.633802 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:09 crc kubenswrapper[4989]: I1006 10:29:09.442578 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:09 crc kubenswrapper[4989]: I1006 10:29:09.442637 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:09 crc kubenswrapper[4989]: I1006 10:29:09.503457 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:09 crc kubenswrapper[4989]: I1006 10:29:09.632566 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:11 crc kubenswrapper[4989]: I1006 10:29:11.294327 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:29:11 crc kubenswrapper[4989]: I1006 10:29:11.295148 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l9k5q" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="registry-server" containerID="cri-o://7c030879e5b59543b1c061f84727b2f73cba0187d1786487a64d1b3e7fcec891" gracePeriod=2 Oct 06 10:29:11 crc kubenswrapper[4989]: I1006 10:29:11.592625 4989 generic.go:334] "Generic (PLEG): container finished" podID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerID="7c030879e5b59543b1c061f84727b2f73cba0187d1786487a64d1b3e7fcec891" exitCode=0 Oct 06 10:29:11 crc kubenswrapper[4989]: I1006 10:29:11.592694 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerDied","Data":"7c030879e5b59543b1c061f84727b2f73cba0187d1786487a64d1b3e7fcec891"} Oct 06 10:29:11 crc kubenswrapper[4989]: I1006 10:29:11.842034 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.015262 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content\") pod \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.015455 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities\") pod \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.015595 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wx6h\" (UniqueName: \"kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h\") pod \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\" (UID: \"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a\") " Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.016162 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities" (OuterVolumeSpecName: "utilities") pod "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" (UID: "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.030859 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h" (OuterVolumeSpecName: "kube-api-access-2wx6h") pod "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" (UID: "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a"). InnerVolumeSpecName "kube-api-access-2wx6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.089798 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" (UID: "bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.117999 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wx6h\" (UniqueName: \"kubernetes.io/projected/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-kube-api-access-2wx6h\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.118032 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.118042 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.606976 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9k5q" event={"ID":"bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a","Type":"ContainerDied","Data":"cb3faf0e8de663063b46c981946c5cc016c21fc2d38390e8fca0b1fd42542764"} Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.607053 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9k5q" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.607092 4989 scope.go:117] "RemoveContainer" containerID="7c030879e5b59543b1c061f84727b2f73cba0187d1786487a64d1b3e7fcec891" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.637455 4989 scope.go:117] "RemoveContainer" containerID="8ff660243f1660d251d6b26304586e1d0b2a54162f24006a60bec92cd9a9c6fa" Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.655414 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.666857 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l9k5q"] Oct 06 10:29:12 crc kubenswrapper[4989]: I1006 10:29:12.667230 4989 scope.go:117] "RemoveContainer" containerID="60984051408d41785419b7e3647ba7fe66c5e69227d778b64e82a9e5d79e3c78" Oct 06 10:29:13 crc kubenswrapper[4989]: I1006 10:29:13.950892 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" path="/var/lib/kubelet/pods/bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a/volumes" Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.304843 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.305561 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-drjhz" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="registry-server" containerID="cri-o://04f9be063e8fa447d08b7674e22491d12dc35f5866e94186d584a34b72407c2f" gracePeriod=2 Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.644154 4989 generic.go:334] "Generic (PLEG): container finished" podID="7d7b2894-777e-4bbe-a290-a27012f07476" containerID="04f9be063e8fa447d08b7674e22491d12dc35f5866e94186d584a34b72407c2f" exitCode=0 Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.644191 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerDied","Data":"04f9be063e8fa447d08b7674e22491d12dc35f5866e94186d584a34b72407c2f"} Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.810633 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.990255 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities\") pod \"7d7b2894-777e-4bbe-a290-a27012f07476\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.990548 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content\") pod \"7d7b2894-777e-4bbe-a290-a27012f07476\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.990632 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8fj6\" (UniqueName: \"kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6\") pod \"7d7b2894-777e-4bbe-a290-a27012f07476\" (UID: \"7d7b2894-777e-4bbe-a290-a27012f07476\") " Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.991267 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities" (OuterVolumeSpecName: "utilities") pod "7d7b2894-777e-4bbe-a290-a27012f07476" (UID: "7d7b2894-777e-4bbe-a290-a27012f07476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:29:14 crc kubenswrapper[4989]: I1006 10:29:14.995685 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6" (OuterVolumeSpecName: "kube-api-access-z8fj6") pod "7d7b2894-777e-4bbe-a290-a27012f07476" (UID: "7d7b2894-777e-4bbe-a290-a27012f07476"). InnerVolumeSpecName "kube-api-access-z8fj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.004583 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d7b2894-777e-4bbe-a290-a27012f07476" (UID: "7d7b2894-777e-4bbe-a290-a27012f07476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.093756 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.093802 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d7b2894-777e-4bbe-a290-a27012f07476-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.093815 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8fj6\" (UniqueName: \"kubernetes.io/projected/7d7b2894-777e-4bbe-a290-a27012f07476-kube-api-access-z8fj6\") on node \"crc\" DevicePath \"\"" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.661412 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drjhz" event={"ID":"7d7b2894-777e-4bbe-a290-a27012f07476","Type":"ContainerDied","Data":"530295e7454d2f21667c8fda38ed16989a1c3b65b519abb3c84bb2b32786169b"} Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.661549 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drjhz" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.661803 4989 scope.go:117] "RemoveContainer" containerID="04f9be063e8fa447d08b7674e22491d12dc35f5866e94186d584a34b72407c2f" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.686259 4989 scope.go:117] "RemoveContainer" containerID="2c79f94264f32d153dd7f4ec4bc3429f04f6bd6ea8e59d40d4b35dc5795d4b4a" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.706897 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.714521 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-drjhz"] Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.727345 4989 scope.go:117] "RemoveContainer" containerID="a514d0159a6b6a3d7d1e810bb9abfbd20bc6c7f288137ec12e2027b768dbba10" Oct 06 10:29:15 crc kubenswrapper[4989]: I1006 10:29:15.953174 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" path="/var/lib/kubelet/pods/7d7b2894-777e-4bbe-a290-a27012f07476/volumes" Oct 06 10:29:16 crc kubenswrapper[4989]: I1006 10:29:16.049101 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-lqqtw"] Oct 06 10:29:16 crc kubenswrapper[4989]: I1006 10:29:16.061236 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-lqqtw"] Oct 06 10:29:17 crc kubenswrapper[4989]: I1006 10:29:17.956806 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="987d1d1a-c7a5-4122-a2eb-86b811863dd5" path="/var/lib/kubelet/pods/987d1d1a-c7a5-4122-a2eb-86b811863dd5/volumes" Oct 06 10:29:25 crc kubenswrapper[4989]: I1006 10:29:25.410279 4989 scope.go:117] "RemoveContainer" containerID="b19d4ca0d4149c76229263d7308475c7e74976d55ecba36b6f416803dc08b9bc" Oct 06 10:29:29 crc kubenswrapper[4989]: I1006 10:29:29.031275 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-7f15-account-create-nhcgx"] Oct 06 10:29:29 crc kubenswrapper[4989]: I1006 10:29:29.043778 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-7f15-account-create-nhcgx"] Oct 06 10:29:29 crc kubenswrapper[4989]: I1006 10:29:29.963544 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4080dd-79fe-4dcd-88f9-80ddb79ab629" path="/var/lib/kubelet/pods/ad4080dd-79fe-4dcd-88f9-80ddb79ab629/volumes" Oct 06 10:29:35 crc kubenswrapper[4989]: I1006 10:29:35.041248 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-kxqsb"] Oct 06 10:29:35 crc kubenswrapper[4989]: I1006 10:29:35.052053 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-kxqsb"] Oct 06 10:29:35 crc kubenswrapper[4989]: I1006 10:29:35.960850 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e52fd09-c241-4aa6-b94d-5e0a653c7f10" path="/var/lib/kubelet/pods/2e52fd09-c241-4aa6-b94d-5e0a653c7f10/volumes" Oct 06 10:29:46 crc kubenswrapper[4989]: I1006 10:29:46.028415 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-4547-account-create-wz8wn"] Oct 06 10:29:46 crc kubenswrapper[4989]: I1006 10:29:46.039927 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-4547-account-create-wz8wn"] Oct 06 10:29:47 crc kubenswrapper[4989]: I1006 10:29:47.961997 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1513a90e-5b01-4b75-a007-97aaf1daa3d7" path="/var/lib/kubelet/pods/1513a90e-5b01-4b75-a007-97aaf1daa3d7/volumes" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.182032 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv"] Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183087 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="extract-content" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183102 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="extract-content" Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183127 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183132 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183145 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="extract-content" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183152 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="extract-content" Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183160 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="extract-utilities" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183166 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="extract-utilities" Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183178 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183186 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: E1006 10:30:00.183210 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="extract-utilities" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183218 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="extract-utilities" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183458 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdbf2783-5db7-4544-8d2d-0b3a8ec7c61a" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.183474 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d7b2894-777e-4bbe-a290-a27012f07476" containerName="registry-server" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.184378 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.186859 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.186982 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.203517 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv"] Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.302628 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.302762 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.302997 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7sq\" (UniqueName: \"kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.405124 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7sq\" (UniqueName: \"kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.405369 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.405441 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.407081 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.419168 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.438377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7sq\" (UniqueName: \"kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq\") pod \"collect-profiles-29329110-n6xqv\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.508816 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:00 crc kubenswrapper[4989]: I1006 10:30:00.967123 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv"] Oct 06 10:30:01 crc kubenswrapper[4989]: I1006 10:30:01.173641 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" event={"ID":"841bea16-2f0a-4327-b3b9-44c16d73213a","Type":"ContainerStarted","Data":"42e08aa8e28abffdcc56ed8e4b6ad8d2837ce327ee55d8323aac87b9b82234c0"} Oct 06 10:30:01 crc kubenswrapper[4989]: I1006 10:30:01.173976 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" event={"ID":"841bea16-2f0a-4327-b3b9-44c16d73213a","Type":"ContainerStarted","Data":"763fddf6131a4bf2e0bdfcfa67885edb6e516e9f42e0dbea53dd37529c90d3dd"} Oct 06 10:30:01 crc kubenswrapper[4989]: I1006 10:30:01.199760 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" podStartSLOduration=1.199734563 podStartE2EDuration="1.199734563s" podCreationTimestamp="2025-10-06 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 10:30:01.190469006 +0000 UTC m=+6651.980494586" watchObservedRunningTime="2025-10-06 10:30:01.199734563 +0000 UTC m=+6651.989760163" Oct 06 10:30:02 crc kubenswrapper[4989]: I1006 10:30:02.188605 4989 generic.go:334] "Generic (PLEG): container finished" podID="841bea16-2f0a-4327-b3b9-44c16d73213a" containerID="42e08aa8e28abffdcc56ed8e4b6ad8d2837ce327ee55d8323aac87b9b82234c0" exitCode=0 Oct 06 10:30:02 crc kubenswrapper[4989]: I1006 10:30:02.189125 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" event={"ID":"841bea16-2f0a-4327-b3b9-44c16d73213a","Type":"ContainerDied","Data":"42e08aa8e28abffdcc56ed8e4b6ad8d2837ce327ee55d8323aac87b9b82234c0"} Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.631901 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.704552 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume\") pod \"841bea16-2f0a-4327-b3b9-44c16d73213a\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.704774 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume\") pod \"841bea16-2f0a-4327-b3b9-44c16d73213a\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.705016 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7sq\" (UniqueName: \"kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq\") pod \"841bea16-2f0a-4327-b3b9-44c16d73213a\" (UID: \"841bea16-2f0a-4327-b3b9-44c16d73213a\") " Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.705473 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume" (OuterVolumeSpecName: "config-volume") pod "841bea16-2f0a-4327-b3b9-44c16d73213a" (UID: "841bea16-2f0a-4327-b3b9-44c16d73213a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.712455 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq" (OuterVolumeSpecName: "kube-api-access-ch7sq") pod "841bea16-2f0a-4327-b3b9-44c16d73213a" (UID: "841bea16-2f0a-4327-b3b9-44c16d73213a"). InnerVolumeSpecName "kube-api-access-ch7sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.713223 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "841bea16-2f0a-4327-b3b9-44c16d73213a" (UID: "841bea16-2f0a-4327-b3b9-44c16d73213a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.806962 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7sq\" (UniqueName: \"kubernetes.io/projected/841bea16-2f0a-4327-b3b9-44c16d73213a-kube-api-access-ch7sq\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.806991 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/841bea16-2f0a-4327-b3b9-44c16d73213a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:03 crc kubenswrapper[4989]: I1006 10:30:03.807002 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/841bea16-2f0a-4327-b3b9-44c16d73213a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:30:04 crc kubenswrapper[4989]: I1006 10:30:04.212143 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" event={"ID":"841bea16-2f0a-4327-b3b9-44c16d73213a","Type":"ContainerDied","Data":"763fddf6131a4bf2e0bdfcfa67885edb6e516e9f42e0dbea53dd37529c90d3dd"} Oct 06 10:30:04 crc kubenswrapper[4989]: I1006 10:30:04.212406 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763fddf6131a4bf2e0bdfcfa67885edb6e516e9f42e0dbea53dd37529c90d3dd" Oct 06 10:30:04 crc kubenswrapper[4989]: I1006 10:30:04.212271 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv" Oct 06 10:30:04 crc kubenswrapper[4989]: I1006 10:30:04.268143 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s"] Oct 06 10:30:04 crc kubenswrapper[4989]: I1006 10:30:04.275976 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329065-bv75s"] Oct 06 10:30:05 crc kubenswrapper[4989]: I1006 10:30:05.956542 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a" path="/var/lib/kubelet/pods/8c1e3b8f-110a-45d8-8ab1-24d6cddd3c4a/volumes" Oct 06 10:30:25 crc kubenswrapper[4989]: I1006 10:30:25.513052 4989 scope.go:117] "RemoveContainer" containerID="1c8a5cff745212002f5fe26f49aacad0daa715a38f7f7d9956b532869cd6da74" Oct 06 10:30:25 crc kubenswrapper[4989]: I1006 10:30:25.560841 4989 scope.go:117] "RemoveContainer" containerID="dcbbe13bc2e4f01bdd8a95e7d264a2c870dbed5f1fcf2f39737146096d6eb6cf" Oct 06 10:30:25 crc kubenswrapper[4989]: I1006 10:30:25.600763 4989 scope.go:117] "RemoveContainer" containerID="ff0a30952b8f7865acfb4f535750171b9456663c8f65acdcfc46f81cd415a23d" Oct 06 10:30:25 crc kubenswrapper[4989]: I1006 10:30:25.669822 4989 scope.go:117] "RemoveContainer" containerID="2d664d9f2f9a25472a1da0ba5c027a440c085dfbcabfe9de7fde71da65abc7ba" Oct 06 10:30:30 crc kubenswrapper[4989]: I1006 10:30:30.046974 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-jcn7j"] Oct 06 10:30:30 crc kubenswrapper[4989]: I1006 10:30:30.063728 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-jcn7j"] Oct 06 10:30:31 crc kubenswrapper[4989]: I1006 10:30:31.948049 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46711b9f-dd60-4901-9f45-a3301b4ea9c6" path="/var/lib/kubelet/pods/46711b9f-dd60-4901-9f45-a3301b4ea9c6/volumes" Oct 06 10:30:33 crc kubenswrapper[4989]: I1006 10:30:33.934917 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:30:33 crc kubenswrapper[4989]: I1006 10:30:33.935244 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:31:03 crc kubenswrapper[4989]: I1006 10:31:03.935399 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:31:03 crc kubenswrapper[4989]: I1006 10:31:03.935988 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:31:25 crc kubenswrapper[4989]: I1006 10:31:25.796296 4989 scope.go:117] "RemoveContainer" containerID="03853485f39b9e0b613725d24e755ca49c884ebd65435baa89a21f7fb105cc30" Oct 06 10:31:25 crc kubenswrapper[4989]: I1006 10:31:25.859040 4989 scope.go:117] "RemoveContainer" containerID="fa254e9869eb48e481ff7d0cf58ee46418e55d6480f5242b927bde858d1bd328" Oct 06 10:31:33 crc kubenswrapper[4989]: I1006 10:31:33.936022 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:31:33 crc kubenswrapper[4989]: I1006 10:31:33.936644 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:31:33 crc kubenswrapper[4989]: I1006 10:31:33.954619 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:31:33 crc kubenswrapper[4989]: I1006 10:31:33.957005 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:31:33 crc kubenswrapper[4989]: I1006 10:31:33.957071 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" gracePeriod=600 Oct 06 10:31:34 crc kubenswrapper[4989]: E1006 10:31:34.088419 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:31:34 crc kubenswrapper[4989]: I1006 10:31:34.219791 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" exitCode=0 Oct 06 10:31:34 crc kubenswrapper[4989]: I1006 10:31:34.219829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d"} Oct 06 10:31:34 crc kubenswrapper[4989]: I1006 10:31:34.219861 4989 scope.go:117] "RemoveContainer" containerID="3a274f3b598f55c9bce71e894b9ff558bdae3c4de39333d4a1ad6ad026ef8f46" Oct 06 10:31:34 crc kubenswrapper[4989]: I1006 10:31:34.220690 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:31:34 crc kubenswrapper[4989]: E1006 10:31:34.221142 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:31:47 crc kubenswrapper[4989]: I1006 10:31:47.936867 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:31:47 crc kubenswrapper[4989]: E1006 10:31:47.938460 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:00 crc kubenswrapper[4989]: I1006 10:32:00.936588 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:32:00 crc kubenswrapper[4989]: E1006 10:32:00.937684 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:12 crc kubenswrapper[4989]: I1006 10:32:12.937188 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:32:12 crc kubenswrapper[4989]: E1006 10:32:12.937968 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:25 crc kubenswrapper[4989]: I1006 10:32:25.970555 4989 scope.go:117] "RemoveContainer" containerID="c8c182c30571bdc18d84e5b1bbf5775ad0a11f256d31e74afd05092e0476f8a1" Oct 06 10:32:26 crc kubenswrapper[4989]: I1006 10:32:26.037380 4989 scope.go:117] "RemoveContainer" containerID="9e77cc225a442bd453aa46829a0fd51b93198a75e4215367499c6c3ad26c2b9e" Oct 06 10:32:26 crc kubenswrapper[4989]: I1006 10:32:26.087901 4989 scope.go:117] "RemoveContainer" containerID="1950a5dd3aec7172fd86482c62193f94c6a5254c52a32dedeed8e65ecb63ca35" Oct 06 10:32:27 crc kubenswrapper[4989]: I1006 10:32:27.936318 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:32:27 crc kubenswrapper[4989]: E1006 10:32:27.937187 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:40 crc kubenswrapper[4989]: I1006 10:32:40.936022 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:32:40 crc kubenswrapper[4989]: E1006 10:32:40.937066 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:53 crc kubenswrapper[4989]: I1006 10:32:53.937269 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:32:53 crc kubenswrapper[4989]: E1006 10:32:53.938622 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:32:59 crc kubenswrapper[4989]: I1006 10:32:59.038374 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-2284z"] Oct 06 10:32:59 crc kubenswrapper[4989]: I1006 10:32:59.049752 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-2284z"] Oct 06 10:32:59 crc kubenswrapper[4989]: I1006 10:32:59.951984 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e70b45b-b8f8-43e9-aec7-fc621835edf1" path="/var/lib/kubelet/pods/1e70b45b-b8f8-43e9-aec7-fc621835edf1/volumes" Oct 06 10:33:07 crc kubenswrapper[4989]: I1006 10:33:07.937126 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:33:07 crc kubenswrapper[4989]: E1006 10:33:07.938555 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:33:10 crc kubenswrapper[4989]: I1006 10:33:10.042873 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-e57b-account-create-sfjbt"] Oct 06 10:33:10 crc kubenswrapper[4989]: I1006 10:33:10.053129 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-e57b-account-create-sfjbt"] Oct 06 10:33:11 crc kubenswrapper[4989]: I1006 10:33:11.949325 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1ec8cad-e2b5-4839-a98c-e3da3169daac" path="/var/lib/kubelet/pods/f1ec8cad-e2b5-4839-a98c-e3da3169daac/volumes" Oct 06 10:33:18 crc kubenswrapper[4989]: I1006 10:33:18.936237 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:33:18 crc kubenswrapper[4989]: E1006 10:33:18.937056 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:33:24 crc kubenswrapper[4989]: I1006 10:33:24.038905 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-28p8d"] Oct 06 10:33:24 crc kubenswrapper[4989]: I1006 10:33:24.049447 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-28p8d"] Oct 06 10:33:25 crc kubenswrapper[4989]: I1006 10:33:25.975609 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a261e667-22d6-4964-a078-425699474660" path="/var/lib/kubelet/pods/a261e667-22d6-4964-a078-425699474660/volumes" Oct 06 10:33:26 crc kubenswrapper[4989]: I1006 10:33:26.158709 4989 scope.go:117] "RemoveContainer" containerID="25dde054c63c2797e3961c0a33dcdf13ccd390c7dd61497e77bd47cc798c7599" Oct 06 10:33:26 crc kubenswrapper[4989]: I1006 10:33:26.185565 4989 scope.go:117] "RemoveContainer" containerID="39b2f69f1486873ad1cf5049bfdf8eac684e9d630f110f5dd9a25bd26badd408" Oct 06 10:33:26 crc kubenswrapper[4989]: I1006 10:33:26.234836 4989 scope.go:117] "RemoveContainer" containerID="46a0d01d6b704a92bf454c1fc3c4256dab49702466c85e57a6d662691a016e60" Oct 06 10:33:29 crc kubenswrapper[4989]: I1006 10:33:29.942282 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:33:29 crc kubenswrapper[4989]: E1006 10:33:29.943055 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:33:43 crc kubenswrapper[4989]: I1006 10:33:43.937459 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:33:43 crc kubenswrapper[4989]: E1006 10:33:43.938467 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:33:56 crc kubenswrapper[4989]: I1006 10:33:56.936367 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:33:56 crc kubenswrapper[4989]: E1006 10:33:56.937061 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:34:09 crc kubenswrapper[4989]: I1006 10:34:09.952080 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:34:09 crc kubenswrapper[4989]: E1006 10:34:09.953336 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:34:20 crc kubenswrapper[4989]: I1006 10:34:20.935645 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:34:20 crc kubenswrapper[4989]: E1006 10:34:20.936456 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:34:33 crc kubenswrapper[4989]: I1006 10:34:33.937012 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:34:33 crc kubenswrapper[4989]: E1006 10:34:33.938169 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:34:44 crc kubenswrapper[4989]: I1006 10:34:44.937712 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:34:44 crc kubenswrapper[4989]: E1006 10:34:44.938558 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:34:59 crc kubenswrapper[4989]: I1006 10:34:59.965911 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:34:59 crc kubenswrapper[4989]: E1006 10:34:59.967118 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.525352 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:09 crc kubenswrapper[4989]: E1006 10:35:09.527703 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841bea16-2f0a-4327-b3b9-44c16d73213a" containerName="collect-profiles" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.527732 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="841bea16-2f0a-4327-b3b9-44c16d73213a" containerName="collect-profiles" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.530837 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="841bea16-2f0a-4327-b3b9-44c16d73213a" containerName="collect-profiles" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.535611 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.535815 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.698818 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmnmt\" (UniqueName: \"kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.699030 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.699090 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.800811 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.801189 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.801374 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmnmt\" (UniqueName: \"kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.801443 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.801551 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.823820 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmnmt\" (UniqueName: \"kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt\") pod \"redhat-operators-7ndk6\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:09 crc kubenswrapper[4989]: I1006 10:35:09.864314 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:10 crc kubenswrapper[4989]: I1006 10:35:10.406866 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:10 crc kubenswrapper[4989]: I1006 10:35:10.477019 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerStarted","Data":"27ced51a6965caed1179d792bf4f3a7dfbf5b254a9e12278b9e53eba1e264ce2"} Oct 06 10:35:11 crc kubenswrapper[4989]: I1006 10:35:11.495577 4989 generic.go:334] "Generic (PLEG): container finished" podID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerID="178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d" exitCode=0 Oct 06 10:35:11 crc kubenswrapper[4989]: I1006 10:35:11.495825 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerDied","Data":"178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d"} Oct 06 10:35:11 crc kubenswrapper[4989]: I1006 10:35:11.501455 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:35:13 crc kubenswrapper[4989]: I1006 10:35:13.523853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerStarted","Data":"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947"} Oct 06 10:35:13 crc kubenswrapper[4989]: I1006 10:35:13.936357 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:35:13 crc kubenswrapper[4989]: E1006 10:35:13.936608 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:35:16 crc kubenswrapper[4989]: I1006 10:35:16.558319 4989 generic.go:334] "Generic (PLEG): container finished" podID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerID="70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947" exitCode=0 Oct 06 10:35:16 crc kubenswrapper[4989]: I1006 10:35:16.558610 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerDied","Data":"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947"} Oct 06 10:35:18 crc kubenswrapper[4989]: I1006 10:35:18.581012 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerStarted","Data":"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725"} Oct 06 10:35:18 crc kubenswrapper[4989]: I1006 10:35:18.601629 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7ndk6" podStartSLOduration=3.650449207 podStartE2EDuration="9.601609127s" podCreationTimestamp="2025-10-06 10:35:09 +0000 UTC" firstStartedPulling="2025-10-06 10:35:11.500865411 +0000 UTC m=+6962.290891031" lastFinishedPulling="2025-10-06 10:35:17.452025371 +0000 UTC m=+6968.242050951" observedRunningTime="2025-10-06 10:35:18.595916714 +0000 UTC m=+6969.385942314" watchObservedRunningTime="2025-10-06 10:35:18.601609127 +0000 UTC m=+6969.391634727" Oct 06 10:35:19 crc kubenswrapper[4989]: I1006 10:35:19.865106 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:19 crc kubenswrapper[4989]: I1006 10:35:19.865432 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:20 crc kubenswrapper[4989]: I1006 10:35:20.921542 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7ndk6" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="registry-server" probeResult="failure" output=< Oct 06 10:35:20 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 10:35:20 crc kubenswrapper[4989]: > Oct 06 10:35:24 crc kubenswrapper[4989]: I1006 10:35:24.937250 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:35:24 crc kubenswrapper[4989]: E1006 10:35:24.937873 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:35:29 crc kubenswrapper[4989]: I1006 10:35:29.962774 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:30 crc kubenswrapper[4989]: I1006 10:35:30.028780 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:30 crc kubenswrapper[4989]: I1006 10:35:30.048820 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-xngpw"] Oct 06 10:35:30 crc kubenswrapper[4989]: I1006 10:35:30.061122 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-xngpw"] Oct 06 10:35:30 crc kubenswrapper[4989]: I1006 10:35:30.216244 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:31 crc kubenswrapper[4989]: I1006 10:35:31.711732 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7ndk6" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="registry-server" containerID="cri-o://c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725" gracePeriod=2 Oct 06 10:35:31 crc kubenswrapper[4989]: I1006 10:35:31.972370 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f45596-8c3f-42fb-90db-43dfc5949438" path="/var/lib/kubelet/pods/75f45596-8c3f-42fb-90db-43dfc5949438/volumes" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.253094 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.352231 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities\") pod \"cafdf9d8-f009-4b50-b402-973ec7c16cea\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.352500 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content\") pod \"cafdf9d8-f009-4b50-b402-973ec7c16cea\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.352730 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmnmt\" (UniqueName: \"kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt\") pod \"cafdf9d8-f009-4b50-b402-973ec7c16cea\" (UID: \"cafdf9d8-f009-4b50-b402-973ec7c16cea\") " Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.352858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities" (OuterVolumeSpecName: "utilities") pod "cafdf9d8-f009-4b50-b402-973ec7c16cea" (UID: "cafdf9d8-f009-4b50-b402-973ec7c16cea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.353622 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.358921 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt" (OuterVolumeSpecName: "kube-api-access-zmnmt") pod "cafdf9d8-f009-4b50-b402-973ec7c16cea" (UID: "cafdf9d8-f009-4b50-b402-973ec7c16cea"). InnerVolumeSpecName "kube-api-access-zmnmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.437530 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cafdf9d8-f009-4b50-b402-973ec7c16cea" (UID: "cafdf9d8-f009-4b50-b402-973ec7c16cea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.455974 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmnmt\" (UniqueName: \"kubernetes.io/projected/cafdf9d8-f009-4b50-b402-973ec7c16cea-kube-api-access-zmnmt\") on node \"crc\" DevicePath \"\"" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.456006 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cafdf9d8-f009-4b50-b402-973ec7c16cea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.726062 4989 generic.go:334] "Generic (PLEG): container finished" podID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerID="c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725" exitCode=0 Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.726147 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerDied","Data":"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725"} Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.726252 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7ndk6" event={"ID":"cafdf9d8-f009-4b50-b402-973ec7c16cea","Type":"ContainerDied","Data":"27ced51a6965caed1179d792bf4f3a7dfbf5b254a9e12278b9e53eba1e264ce2"} Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.726335 4989 scope.go:117] "RemoveContainer" containerID="c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.726166 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7ndk6" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.755454 4989 scope.go:117] "RemoveContainer" containerID="70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.783920 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.797947 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7ndk6"] Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.810190 4989 scope.go:117] "RemoveContainer" containerID="178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.876082 4989 scope.go:117] "RemoveContainer" containerID="c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725" Oct 06 10:35:32 crc kubenswrapper[4989]: E1006 10:35:32.877994 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725\": container with ID starting with c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725 not found: ID does not exist" containerID="c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.878038 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725"} err="failed to get container status \"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725\": rpc error: code = NotFound desc = could not find container \"c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725\": container with ID starting with c17928677a6ba8585e10aa6335346c6a962c4a6473684f7b006fc5fe29ccd725 not found: ID does not exist" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.878065 4989 scope.go:117] "RemoveContainer" containerID="70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947" Oct 06 10:35:32 crc kubenswrapper[4989]: E1006 10:35:32.878511 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947\": container with ID starting with 70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947 not found: ID does not exist" containerID="70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.878558 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947"} err="failed to get container status \"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947\": rpc error: code = NotFound desc = could not find container \"70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947\": container with ID starting with 70ebc537676ea55cdca2676857c0dd1288d8ec1e66e5c307acf27045674da947 not found: ID does not exist" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.878582 4989 scope.go:117] "RemoveContainer" containerID="178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d" Oct 06 10:35:32 crc kubenswrapper[4989]: E1006 10:35:32.878934 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d\": container with ID starting with 178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d not found: ID does not exist" containerID="178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d" Oct 06 10:35:32 crc kubenswrapper[4989]: I1006 10:35:32.878979 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d"} err="failed to get container status \"178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d\": rpc error: code = NotFound desc = could not find container \"178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d\": container with ID starting with 178058dc8f187940c6f37daf335685a03fe21bf332109135666a866373c2dc6d not found: ID does not exist" Oct 06 10:35:33 crc kubenswrapper[4989]: I1006 10:35:33.954393 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" path="/var/lib/kubelet/pods/cafdf9d8-f009-4b50-b402-973ec7c16cea/volumes" Oct 06 10:35:38 crc kubenswrapper[4989]: I1006 10:35:38.935938 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:35:38 crc kubenswrapper[4989]: E1006 10:35:38.936365 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:35:40 crc kubenswrapper[4989]: I1006 10:35:40.043917 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-7710-account-create-hjtcv"] Oct 06 10:35:40 crc kubenswrapper[4989]: I1006 10:35:40.051384 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-7710-account-create-hjtcv"] Oct 06 10:35:41 crc kubenswrapper[4989]: I1006 10:35:41.952194 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ae70841-1f24-4bc0-8b0e-3614cfa21862" path="/var/lib/kubelet/pods/3ae70841-1f24-4bc0-8b0e-3614cfa21862/volumes" Oct 06 10:35:51 crc kubenswrapper[4989]: I1006 10:35:51.025229 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-jj6ql"] Oct 06 10:35:51 crc kubenswrapper[4989]: I1006 10:35:51.033528 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-jj6ql"] Oct 06 10:35:51 crc kubenswrapper[4989]: I1006 10:35:51.949793 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e256cbde-f7d0-4065-ae22-c57ab646e143" path="/var/lib/kubelet/pods/e256cbde-f7d0-4065-ae22-c57ab646e143/volumes" Oct 06 10:35:53 crc kubenswrapper[4989]: I1006 10:35:53.935965 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:35:53 crc kubenswrapper[4989]: E1006 10:35:53.936559 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.523370 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:05 crc kubenswrapper[4989]: E1006 10:36:05.524580 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="extract-utilities" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.524602 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="extract-utilities" Oct 06 10:36:05 crc kubenswrapper[4989]: E1006 10:36:05.524685 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="extract-content" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.524696 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="extract-content" Oct 06 10:36:05 crc kubenswrapper[4989]: E1006 10:36:05.524732 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="registry-server" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.524743 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="registry-server" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.525119 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="cafdf9d8-f009-4b50-b402-973ec7c16cea" containerName="registry-server" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.527383 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.535487 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.698176 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sqqs\" (UniqueName: \"kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.698477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.698616 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.801273 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sqqs\" (UniqueName: \"kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.801693 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.801894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.802201 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.802417 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.831397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sqqs\" (UniqueName: \"kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs\") pod \"certified-operators-4jp2n\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:05 crc kubenswrapper[4989]: I1006 10:36:05.859104 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:06 crc kubenswrapper[4989]: I1006 10:36:06.410988 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:07 crc kubenswrapper[4989]: I1006 10:36:07.114778 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerID="49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350" exitCode=0 Oct 06 10:36:07 crc kubenswrapper[4989]: I1006 10:36:07.114864 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerDied","Data":"49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350"} Oct 06 10:36:07 crc kubenswrapper[4989]: I1006 10:36:07.115161 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerStarted","Data":"1a2c2aa4277ea68fbe24e1219e42c11ed37cd00df715d9c7d2c89b60942b0071"} Oct 06 10:36:07 crc kubenswrapper[4989]: I1006 10:36:07.938344 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:36:07 crc kubenswrapper[4989]: E1006 10:36:07.939317 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:36:08 crc kubenswrapper[4989]: I1006 10:36:08.126853 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerStarted","Data":"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6"} Oct 06 10:36:10 crc kubenswrapper[4989]: I1006 10:36:10.156485 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerID="65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6" exitCode=0 Oct 06 10:36:10 crc kubenswrapper[4989]: I1006 10:36:10.156876 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerDied","Data":"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6"} Oct 06 10:36:12 crc kubenswrapper[4989]: I1006 10:36:12.042904 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-449sh"] Oct 06 10:36:12 crc kubenswrapper[4989]: I1006 10:36:12.052206 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-449sh"] Oct 06 10:36:12 crc kubenswrapper[4989]: I1006 10:36:12.184867 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerStarted","Data":"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256"} Oct 06 10:36:12 crc kubenswrapper[4989]: I1006 10:36:12.207076 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jp2n" podStartSLOduration=3.114861127 podStartE2EDuration="7.207058454s" podCreationTimestamp="2025-10-06 10:36:05 +0000 UTC" firstStartedPulling="2025-10-06 10:36:07.11800368 +0000 UTC m=+7017.908029260" lastFinishedPulling="2025-10-06 10:36:11.210201007 +0000 UTC m=+7022.000226587" observedRunningTime="2025-10-06 10:36:12.199587951 +0000 UTC m=+7022.989613531" watchObservedRunningTime="2025-10-06 10:36:12.207058454 +0000 UTC m=+7022.997084034" Oct 06 10:36:13 crc kubenswrapper[4989]: I1006 10:36:13.953587 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eac29ac-df99-4009-9847-6c38dd9da407" path="/var/lib/kubelet/pods/2eac29ac-df99-4009-9847-6c38dd9da407/volumes" Oct 06 10:36:15 crc kubenswrapper[4989]: I1006 10:36:15.859638 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:15 crc kubenswrapper[4989]: I1006 10:36:15.859719 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:15 crc kubenswrapper[4989]: I1006 10:36:15.912003 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:16 crc kubenswrapper[4989]: I1006 10:36:16.269490 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:16 crc kubenswrapper[4989]: I1006 10:36:16.320180 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.237549 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4jp2n" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="registry-server" containerID="cri-o://2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256" gracePeriod=2 Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.672226 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.854891 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities\") pod \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.855216 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content\") pod \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.855412 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sqqs\" (UniqueName: \"kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs\") pod \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\" (UID: \"fa0374b3-b8c4-4452-9bbd-23210acc5a57\") " Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.855857 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities" (OuterVolumeSpecName: "utilities") pod "fa0374b3-b8c4-4452-9bbd-23210acc5a57" (UID: "fa0374b3-b8c4-4452-9bbd-23210acc5a57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.860813 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.866678 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs" (OuterVolumeSpecName: "kube-api-access-9sqqs") pod "fa0374b3-b8c4-4452-9bbd-23210acc5a57" (UID: "fa0374b3-b8c4-4452-9bbd-23210acc5a57"). InnerVolumeSpecName "kube-api-access-9sqqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.908360 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa0374b3-b8c4-4452-9bbd-23210acc5a57" (UID: "fa0374b3-b8c4-4452-9bbd-23210acc5a57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.962319 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sqqs\" (UniqueName: \"kubernetes.io/projected/fa0374b3-b8c4-4452-9bbd-23210acc5a57-kube-api-access-9sqqs\") on node \"crc\" DevicePath \"\"" Oct 06 10:36:18 crc kubenswrapper[4989]: I1006 10:36:18.962601 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa0374b3-b8c4-4452-9bbd-23210acc5a57-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.256836 4989 generic.go:334] "Generic (PLEG): container finished" podID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerID="2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256" exitCode=0 Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.256886 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerDied","Data":"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256"} Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.256900 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jp2n" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.256916 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jp2n" event={"ID":"fa0374b3-b8c4-4452-9bbd-23210acc5a57","Type":"ContainerDied","Data":"1a2c2aa4277ea68fbe24e1219e42c11ed37cd00df715d9c7d2c89b60942b0071"} Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.256967 4989 scope.go:117] "RemoveContainer" containerID="2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.292918 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.299777 4989 scope.go:117] "RemoveContainer" containerID="65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.301497 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4jp2n"] Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.320784 4989 scope.go:117] "RemoveContainer" containerID="49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.382893 4989 scope.go:117] "RemoveContainer" containerID="2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256" Oct 06 10:36:19 crc kubenswrapper[4989]: E1006 10:36:19.383615 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256\": container with ID starting with 2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256 not found: ID does not exist" containerID="2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.383684 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256"} err="failed to get container status \"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256\": rpc error: code = NotFound desc = could not find container \"2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256\": container with ID starting with 2d6fd90cd758eed23152dd3b0a8526d851a9450f31144caabd5a0050b0ac8256 not found: ID does not exist" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.383710 4989 scope.go:117] "RemoveContainer" containerID="65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6" Oct 06 10:36:19 crc kubenswrapper[4989]: E1006 10:36:19.384261 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6\": container with ID starting with 65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6 not found: ID does not exist" containerID="65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.384297 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6"} err="failed to get container status \"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6\": rpc error: code = NotFound desc = could not find container \"65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6\": container with ID starting with 65ebb758e3b2c2c60b3a0888510f1744655672e47eea3508dafb072bea5d2af6 not found: ID does not exist" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.384317 4989 scope.go:117] "RemoveContainer" containerID="49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350" Oct 06 10:36:19 crc kubenswrapper[4989]: E1006 10:36:19.385269 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350\": container with ID starting with 49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350 not found: ID does not exist" containerID="49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.385307 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350"} err="failed to get container status \"49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350\": rpc error: code = NotFound desc = could not find container \"49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350\": container with ID starting with 49341f1eb898b18999d91227a61daa925fe56df0a9f568e60e45719ae6174350 not found: ID does not exist" Oct 06 10:36:19 crc kubenswrapper[4989]: I1006 10:36:19.956364 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" path="/var/lib/kubelet/pods/fa0374b3-b8c4-4452-9bbd-23210acc5a57/volumes" Oct 06 10:36:21 crc kubenswrapper[4989]: I1006 10:36:21.935886 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:36:21 crc kubenswrapper[4989]: E1006 10:36:21.936564 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:36:22 crc kubenswrapper[4989]: I1006 10:36:22.031873 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-fe32-account-create-kgsdv"] Oct 06 10:36:22 crc kubenswrapper[4989]: I1006 10:36:22.039583 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-fe32-account-create-kgsdv"] Oct 06 10:36:23 crc kubenswrapper[4989]: I1006 10:36:23.964029 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f6a5693-a77c-4f1b-81b9-643a410095cf" path="/var/lib/kubelet/pods/7f6a5693-a77c-4f1b-81b9-643a410095cf/volumes" Oct 06 10:36:26 crc kubenswrapper[4989]: I1006 10:36:26.398856 4989 scope.go:117] "RemoveContainer" containerID="6a947d5bead38a90aca3a239cae3d9cdd91335945936f0d8c91dd6926b1bbcb9" Oct 06 10:36:26 crc kubenswrapper[4989]: I1006 10:36:26.435162 4989 scope.go:117] "RemoveContainer" containerID="32e630257eacbe33dae883b0fc2e7149d94dea04439ee6f8f3f1fef7ba86bf1b" Oct 06 10:36:26 crc kubenswrapper[4989]: I1006 10:36:26.500255 4989 scope.go:117] "RemoveContainer" containerID="c46db30270c5f31c350adcc443c8f6f5ebdff1c6a4bb0d839e407143a9ef919d" Oct 06 10:36:26 crc kubenswrapper[4989]: I1006 10:36:26.537953 4989 scope.go:117] "RemoveContainer" containerID="280af0cdfc131d95d6d757b1d19dec581cee9e1b3078f04397db60079b677112" Oct 06 10:36:26 crc kubenswrapper[4989]: I1006 10:36:26.588413 4989 scope.go:117] "RemoveContainer" containerID="0983db184f561963ee77c819139824b8befeae52e528a4e79f24cf473261a92d" Oct 06 10:36:33 crc kubenswrapper[4989]: I1006 10:36:33.937511 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:36:33 crc kubenswrapper[4989]: E1006 10:36:33.938705 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:36:35 crc kubenswrapper[4989]: I1006 10:36:35.047720 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-7tsmz"] Oct 06 10:36:35 crc kubenswrapper[4989]: I1006 10:36:35.062869 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-7tsmz"] Oct 06 10:36:35 crc kubenswrapper[4989]: I1006 10:36:35.964914 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2be20a5-53e0-4fc6-a4d0-29206fbf5676" path="/var/lib/kubelet/pods/b2be20a5-53e0-4fc6-a4d0-29206fbf5676/volumes" Oct 06 10:36:47 crc kubenswrapper[4989]: I1006 10:36:47.936288 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:36:48 crc kubenswrapper[4989]: I1006 10:36:48.625174 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7"} Oct 06 10:37:26 crc kubenswrapper[4989]: I1006 10:37:26.854480 4989 scope.go:117] "RemoveContainer" containerID="51e8997cfb34cbc8614a653d46de4ece77c1dd218cdaf51c108214506a3c35c8" Oct 06 10:39:03 crc kubenswrapper[4989]: I1006 10:39:03.935822 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:39:03 crc kubenswrapper[4989]: I1006 10:39:03.936305 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:39:09 crc kubenswrapper[4989]: I1006 10:39:09.221725 4989 generic.go:334] "Generic (PLEG): container finished" podID="5e33f7f9-1020-4783-bee8-c3effc5defb2" containerID="dad1ed2fb8f33d2d1617279207f601ce6728e884bb21616ef23fbf4aa733cda2" exitCode=0 Oct 06 10:39:09 crc kubenswrapper[4989]: I1006 10:39:09.221839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" event={"ID":"5e33f7f9-1020-4783-bee8-c3effc5defb2","Type":"ContainerDied","Data":"dad1ed2fb8f33d2d1617279207f601ce6728e884bb21616ef23fbf4aa733cda2"} Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.825569 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.958311 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory\") pod \"5e33f7f9-1020-4783-bee8-c3effc5defb2\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.958455 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key\") pod \"5e33f7f9-1020-4783-bee8-c3effc5defb2\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.958616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gttrl\" (UniqueName: \"kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl\") pod \"5e33f7f9-1020-4783-bee8-c3effc5defb2\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.958693 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph\") pod \"5e33f7f9-1020-4783-bee8-c3effc5defb2\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.959508 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle\") pod \"5e33f7f9-1020-4783-bee8-c3effc5defb2\" (UID: \"5e33f7f9-1020-4783-bee8-c3effc5defb2\") " Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.965917 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "5e33f7f9-1020-4783-bee8-c3effc5defb2" (UID: "5e33f7f9-1020-4783-bee8-c3effc5defb2"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.966321 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl" (OuterVolumeSpecName: "kube-api-access-gttrl") pod "5e33f7f9-1020-4783-bee8-c3effc5defb2" (UID: "5e33f7f9-1020-4783-bee8-c3effc5defb2"). InnerVolumeSpecName "kube-api-access-gttrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.967542 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph" (OuterVolumeSpecName: "ceph") pod "5e33f7f9-1020-4783-bee8-c3effc5defb2" (UID: "5e33f7f9-1020-4783-bee8-c3effc5defb2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.988626 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5e33f7f9-1020-4783-bee8-c3effc5defb2" (UID: "5e33f7f9-1020-4783-bee8-c3effc5defb2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:39:10 crc kubenswrapper[4989]: I1006 10:39:10.998888 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory" (OuterVolumeSpecName: "inventory") pod "5e33f7f9-1020-4783-bee8-c3effc5defb2" (UID: "5e33f7f9-1020-4783-bee8-c3effc5defb2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.062452 4989 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.062503 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.062517 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.062530 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gttrl\" (UniqueName: \"kubernetes.io/projected/5e33f7f9-1020-4783-bee8-c3effc5defb2-kube-api-access-gttrl\") on node \"crc\" DevicePath \"\"" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.062546 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e33f7f9-1020-4783-bee8-c3effc5defb2-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.248025 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" event={"ID":"5e33f7f9-1020-4783-bee8-c3effc5defb2","Type":"ContainerDied","Data":"37f02cbf8490201e2be0586ae92faa0f6b4af24ffe4eaff4e5660eb7351c5580"} Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.248092 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw" Oct 06 10:39:11 crc kubenswrapper[4989]: I1006 10:39:11.248098 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37f02cbf8490201e2be0586ae92faa0f6b4af24ffe4eaff4e5660eb7351c5580" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.481077 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-hdzjv"] Oct 06 10:39:16 crc kubenswrapper[4989]: E1006 10:39:16.482038 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="extract-content" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482051 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="extract-content" Oct 06 10:39:16 crc kubenswrapper[4989]: E1006 10:39:16.482071 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="extract-utilities" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482078 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="extract-utilities" Oct 06 10:39:16 crc kubenswrapper[4989]: E1006 10:39:16.482101 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e33f7f9-1020-4783-bee8-c3effc5defb2" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482109 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e33f7f9-1020-4783-bee8-c3effc5defb2" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 06 10:39:16 crc kubenswrapper[4989]: E1006 10:39:16.482121 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="registry-server" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482126 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="registry-server" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482316 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e33f7f9-1020-4783-bee8-c3effc5defb2" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.482335 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0374b3-b8c4-4452-9bbd-23210acc5a57" containerName="registry-server" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.483128 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.489946 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.490144 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.490412 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.490552 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.495498 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-hdzjv"] Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.580140 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.580199 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.580310 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.580358 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnt7\" (UniqueName: \"kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.580762 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.682423 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.682473 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.682519 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.682555 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnt7\" (UniqueName: \"kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.682651 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.689086 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.689677 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.689972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.690469 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.709077 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnt7\" (UniqueName: \"kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7\") pod \"bootstrap-openstack-openstack-cell1-hdzjv\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:16 crc kubenswrapper[4989]: I1006 10:39:16.802011 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:39:17 crc kubenswrapper[4989]: I1006 10:39:17.562230 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-hdzjv"] Oct 06 10:39:18 crc kubenswrapper[4989]: I1006 10:39:18.338550 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" event={"ID":"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1","Type":"ContainerStarted","Data":"882a0b88b439f2f68eb07137cdd42927ab54367b09f1c35ed496701c176d4311"} Oct 06 10:39:18 crc kubenswrapper[4989]: I1006 10:39:18.338958 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" event={"ID":"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1","Type":"ContainerStarted","Data":"5651b58cbe4ce9b901eea26e19a4fed67585f208287d4377b747f9afb0d2060a"} Oct 06 10:39:33 crc kubenswrapper[4989]: I1006 10:39:33.935668 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:39:33 crc kubenswrapper[4989]: I1006 10:39:33.936203 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:40:03 crc kubenswrapper[4989]: I1006 10:40:03.935237 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:40:03 crc kubenswrapper[4989]: I1006 10:40:03.936346 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:40:03 crc kubenswrapper[4989]: I1006 10:40:03.950974 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:40:03 crc kubenswrapper[4989]: I1006 10:40:03.952021 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:40:03 crc kubenswrapper[4989]: I1006 10:40:03.952108 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7" gracePeriod=600 Oct 06 10:40:04 crc kubenswrapper[4989]: I1006 10:40:04.812878 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7" exitCode=0 Oct 06 10:40:04 crc kubenswrapper[4989]: I1006 10:40:04.813134 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7"} Oct 06 10:40:04 crc kubenswrapper[4989]: I1006 10:40:04.814005 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e"} Oct 06 10:40:04 crc kubenswrapper[4989]: I1006 10:40:04.814313 4989 scope.go:117] "RemoveContainer" containerID="853930f0c80cb72ba4096619e56a4003ee2659ea0bae4402791569487f97f85d" Oct 06 10:40:04 crc kubenswrapper[4989]: I1006 10:40:04.847809 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" podStartSLOduration=48.62556802 podStartE2EDuration="48.847785118s" podCreationTimestamp="2025-10-06 10:39:16 +0000 UTC" firstStartedPulling="2025-10-06 10:39:17.48488167 +0000 UTC m=+7208.274907270" lastFinishedPulling="2025-10-06 10:39:17.707098788 +0000 UTC m=+7208.497124368" observedRunningTime="2025-10-06 10:39:18.380108667 +0000 UTC m=+7209.170134247" watchObservedRunningTime="2025-10-06 10:40:04.847785118 +0000 UTC m=+7255.637810708" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.569338 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.574355 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.585794 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.629574 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.629642 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.630713 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvpt7\" (UniqueName: \"kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.733057 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvpt7\" (UniqueName: \"kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.733193 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.733246 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.733837 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.733845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.758313 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvpt7\" (UniqueName: \"kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7\") pod \"redhat-marketplace-8nczk\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:23 crc kubenswrapper[4989]: I1006 10:40:23.895557 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:24 crc kubenswrapper[4989]: I1006 10:40:24.394754 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:25 crc kubenswrapper[4989]: I1006 10:40:25.026638 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerID="d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e" exitCode=0 Oct 06 10:40:25 crc kubenswrapper[4989]: I1006 10:40:25.026724 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerDied","Data":"d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e"} Oct 06 10:40:25 crc kubenswrapper[4989]: I1006 10:40:25.026966 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerStarted","Data":"fd3d78a91327ea40a806a568ef0934527fc33b8f55432121e93b6b14353e78ca"} Oct 06 10:40:25 crc kubenswrapper[4989]: I1006 10:40:25.028623 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:40:26 crc kubenswrapper[4989]: I1006 10:40:26.038316 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerID="407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b" exitCode=0 Oct 06 10:40:26 crc kubenswrapper[4989]: I1006 10:40:26.038575 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerDied","Data":"407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b"} Oct 06 10:40:27 crc kubenswrapper[4989]: I1006 10:40:27.050007 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerStarted","Data":"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b"} Oct 06 10:40:27 crc kubenswrapper[4989]: I1006 10:40:27.070896 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8nczk" podStartSLOduration=2.488524118 podStartE2EDuration="4.070875387s" podCreationTimestamp="2025-10-06 10:40:23 +0000 UTC" firstStartedPulling="2025-10-06 10:40:25.028395351 +0000 UTC m=+7275.818420931" lastFinishedPulling="2025-10-06 10:40:26.61074661 +0000 UTC m=+7277.400772200" observedRunningTime="2025-10-06 10:40:27.06573385 +0000 UTC m=+7277.855759450" watchObservedRunningTime="2025-10-06 10:40:27.070875387 +0000 UTC m=+7277.860900967" Oct 06 10:40:33 crc kubenswrapper[4989]: I1006 10:40:33.896164 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:33 crc kubenswrapper[4989]: I1006 10:40:33.897039 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:33 crc kubenswrapper[4989]: I1006 10:40:33.958968 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:34 crc kubenswrapper[4989]: I1006 10:40:34.166442 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:34 crc kubenswrapper[4989]: I1006 10:40:34.222746 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.140984 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8nczk" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="registry-server" containerID="cri-o://5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b" gracePeriod=2 Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.670395 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.753767 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content\") pod \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.754036 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities\") pod \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.754103 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvpt7\" (UniqueName: \"kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7\") pod \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\" (UID: \"3bc2714a-733b-4c77-9d4c-3ce79eff0666\") " Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.755570 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities" (OuterVolumeSpecName: "utilities") pod "3bc2714a-733b-4c77-9d4c-3ce79eff0666" (UID: "3bc2714a-733b-4c77-9d4c-3ce79eff0666"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.760786 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7" (OuterVolumeSpecName: "kube-api-access-cvpt7") pod "3bc2714a-733b-4c77-9d4c-3ce79eff0666" (UID: "3bc2714a-733b-4c77-9d4c-3ce79eff0666"). InnerVolumeSpecName "kube-api-access-cvpt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.779374 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bc2714a-733b-4c77-9d4c-3ce79eff0666" (UID: "3bc2714a-733b-4c77-9d4c-3ce79eff0666"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.856914 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.856950 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvpt7\" (UniqueName: \"kubernetes.io/projected/3bc2714a-733b-4c77-9d4c-3ce79eff0666-kube-api-access-cvpt7\") on node \"crc\" DevicePath \"\"" Oct 06 10:40:36 crc kubenswrapper[4989]: I1006 10:40:36.856962 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bc2714a-733b-4c77-9d4c-3ce79eff0666-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.159269 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerID="5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b" exitCode=0 Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.159326 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nczk" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.159354 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerDied","Data":"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b"} Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.159409 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nczk" event={"ID":"3bc2714a-733b-4c77-9d4c-3ce79eff0666","Type":"ContainerDied","Data":"fd3d78a91327ea40a806a568ef0934527fc33b8f55432121e93b6b14353e78ca"} Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.159444 4989 scope.go:117] "RemoveContainer" containerID="5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.208177 4989 scope.go:117] "RemoveContainer" containerID="407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.216831 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.238047 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nczk"] Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.242785 4989 scope.go:117] "RemoveContainer" containerID="d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.296703 4989 scope.go:117] "RemoveContainer" containerID="5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b" Oct 06 10:40:37 crc kubenswrapper[4989]: E1006 10:40:37.298333 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b\": container with ID starting with 5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b not found: ID does not exist" containerID="5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.298364 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b"} err="failed to get container status \"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b\": rpc error: code = NotFound desc = could not find container \"5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b\": container with ID starting with 5518d80b8dac0496c41b3f5bdabbab13032607afa44cc0b0beb97907e2a7694b not found: ID does not exist" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.298384 4989 scope.go:117] "RemoveContainer" containerID="407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b" Oct 06 10:40:37 crc kubenswrapper[4989]: E1006 10:40:37.298940 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b\": container with ID starting with 407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b not found: ID does not exist" containerID="407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.298989 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b"} err="failed to get container status \"407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b\": rpc error: code = NotFound desc = could not find container \"407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b\": container with ID starting with 407b0a8461b0cce7aa4a8ad5fb9659af1d53ebe8a5a774e0e840713baf8b6a2b not found: ID does not exist" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.299024 4989 scope.go:117] "RemoveContainer" containerID="d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e" Oct 06 10:40:37 crc kubenswrapper[4989]: E1006 10:40:37.299434 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e\": container with ID starting with d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e not found: ID does not exist" containerID="d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.299467 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e"} err="failed to get container status \"d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e\": rpc error: code = NotFound desc = could not find container \"d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e\": container with ID starting with d8e11b06d582c9411d175b1371847cd7e14d347d435a12f37707d5fcc570111e not found: ID does not exist" Oct 06 10:40:37 crc kubenswrapper[4989]: I1006 10:40:37.947619 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" path="/var/lib/kubelet/pods/3bc2714a-733b-4c77-9d4c-3ce79eff0666/volumes" Oct 06 10:41:56 crc kubenswrapper[4989]: I1006 10:41:56.176347 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-k5nfm" Oct 06 10:42:21 crc kubenswrapper[4989]: I1006 10:42:21.295007 4989 generic.go:334] "Generic (PLEG): container finished" podID="f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" containerID="882a0b88b439f2f68eb07137cdd42927ab54367b09f1c35ed496701c176d4311" exitCode=0 Oct 06 10:42:21 crc kubenswrapper[4989]: I1006 10:42:21.295576 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" event={"ID":"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1","Type":"ContainerDied","Data":"882a0b88b439f2f68eb07137cdd42927ab54367b09f1c35ed496701c176d4311"} Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.793959 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.889859 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle\") pod \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.890076 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory\") pod \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.890156 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key\") pod \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.890230 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph\") pod \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.890306 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttnt7\" (UniqueName: \"kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7\") pod \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\" (UID: \"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1\") " Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.895719 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" (UID: "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.896863 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7" (OuterVolumeSpecName: "kube-api-access-ttnt7") pod "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" (UID: "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1"). InnerVolumeSpecName "kube-api-access-ttnt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.897783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph" (OuterVolumeSpecName: "ceph") pod "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" (UID: "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.920720 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory" (OuterVolumeSpecName: "inventory") pod "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" (UID: "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.923910 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" (UID: "f088c54d-a955-4cb4-a3af-7c3d1df3c5f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.992574 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.992619 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.992632 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttnt7\" (UniqueName: \"kubernetes.io/projected/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-kube-api-access-ttnt7\") on node \"crc\" DevicePath \"\"" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.992646 4989 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:42:22 crc kubenswrapper[4989]: I1006 10:42:22.992675 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f088c54d-a955-4cb4-a3af-7c3d1df3c5f1-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.337432 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" event={"ID":"f088c54d-a955-4cb4-a3af-7c3d1df3c5f1","Type":"ContainerDied","Data":"5651b58cbe4ce9b901eea26e19a4fed67585f208287d4377b747f9afb0d2060a"} Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.337488 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5651b58cbe4ce9b901eea26e19a4fed67585f208287d4377b747f9afb0d2060a" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.337532 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-hdzjv" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.403683 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-65277"] Oct 06 10:42:23 crc kubenswrapper[4989]: E1006 10:42:23.404119 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="registry-server" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404136 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="registry-server" Oct 06 10:42:23 crc kubenswrapper[4989]: E1006 10:42:23.404145 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" containerName="bootstrap-openstack-openstack-cell1" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404151 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" containerName="bootstrap-openstack-openstack-cell1" Oct 06 10:42:23 crc kubenswrapper[4989]: E1006 10:42:23.404175 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="extract-utilities" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404181 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="extract-utilities" Oct 06 10:42:23 crc kubenswrapper[4989]: E1006 10:42:23.404225 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="extract-content" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404235 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="extract-content" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404490 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="f088c54d-a955-4cb4-a3af-7c3d1df3c5f1" containerName="bootstrap-openstack-openstack-cell1" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.404537 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc2714a-733b-4c77-9d4c-3ce79eff0666" containerName="registry-server" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.405459 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.409378 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.409610 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.409821 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.410015 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.412354 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-65277"] Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.502357 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.502821 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.502884 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.502939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmrn\" (UniqueName: \"kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.604358 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.604517 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.604542 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.604565 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmrn\" (UniqueName: \"kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.609510 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.611003 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.620707 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.623584 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmrn\" (UniqueName: \"kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn\") pod \"download-cache-openstack-openstack-cell1-65277\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:23 crc kubenswrapper[4989]: I1006 10:42:23.747253 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:42:24 crc kubenswrapper[4989]: I1006 10:42:24.297421 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-65277"] Oct 06 10:42:24 crc kubenswrapper[4989]: I1006 10:42:24.349895 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-65277" event={"ID":"8f67e476-79a2-499d-ad1b-142eac5e3681","Type":"ContainerStarted","Data":"160f0be5bae66876732d4f0665e3c7f3f6cfe8e9fc161cea1bf9c1442bcdf049"} Oct 06 10:42:25 crc kubenswrapper[4989]: I1006 10:42:25.360623 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-65277" event={"ID":"8f67e476-79a2-499d-ad1b-142eac5e3681","Type":"ContainerStarted","Data":"8c99b6a24af8675892eb960657745f9560dd91f176390eeb74c373c59582b7f8"} Oct 06 10:42:25 crc kubenswrapper[4989]: I1006 10:42:25.382869 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-65277" podStartSLOduration=2.200874156 podStartE2EDuration="2.382838283s" podCreationTimestamp="2025-10-06 10:42:23 +0000 UTC" firstStartedPulling="2025-10-06 10:42:24.316732757 +0000 UTC m=+7395.106758357" lastFinishedPulling="2025-10-06 10:42:24.498696904 +0000 UTC m=+7395.288722484" observedRunningTime="2025-10-06 10:42:25.37993375 +0000 UTC m=+7396.169959360" watchObservedRunningTime="2025-10-06 10:42:25.382838283 +0000 UTC m=+7396.172863903" Oct 06 10:42:33 crc kubenswrapper[4989]: I1006 10:42:33.935898 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:42:33 crc kubenswrapper[4989]: I1006 10:42:33.936541 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.891799 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.896297 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.925594 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-952lr\" (UniqueName: \"kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.925645 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.925706 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:49 crc kubenswrapper[4989]: I1006 10:42:49.931977 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.027862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-952lr\" (UniqueName: \"kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.028521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.028671 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.029223 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.029588 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.046522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-952lr\" (UniqueName: \"kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr\") pod \"community-operators-dd2m7\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.229435 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:42:50 crc kubenswrapper[4989]: I1006 10:42:50.813960 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:42:51 crc kubenswrapper[4989]: I1006 10:42:51.667277 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerID="b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462" exitCode=0 Oct 06 10:42:51 crc kubenswrapper[4989]: I1006 10:42:51.667702 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerDied","Data":"b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462"} Oct 06 10:42:51 crc kubenswrapper[4989]: I1006 10:42:51.667743 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerStarted","Data":"945cbc97b61ee138c83e796cc983fd55e5b7982c4566ae69f70c47ea2bacd2c5"} Oct 06 10:42:53 crc kubenswrapper[4989]: I1006 10:42:53.698564 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerStarted","Data":"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a"} Oct 06 10:42:55 crc kubenswrapper[4989]: I1006 10:42:55.720490 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerID="d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a" exitCode=0 Oct 06 10:42:55 crc kubenswrapper[4989]: I1006 10:42:55.720566 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerDied","Data":"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a"} Oct 06 10:42:56 crc kubenswrapper[4989]: I1006 10:42:56.733588 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerStarted","Data":"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5"} Oct 06 10:42:56 crc kubenswrapper[4989]: I1006 10:42:56.771199 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dd2m7" podStartSLOduration=3.036318125 podStartE2EDuration="7.771169232s" podCreationTimestamp="2025-10-06 10:42:49 +0000 UTC" firstStartedPulling="2025-10-06 10:42:51.682500731 +0000 UTC m=+7422.472526331" lastFinishedPulling="2025-10-06 10:42:56.417351858 +0000 UTC m=+7427.207377438" observedRunningTime="2025-10-06 10:42:56.76760044 +0000 UTC m=+7427.557626030" watchObservedRunningTime="2025-10-06 10:42:56.771169232 +0000 UTC m=+7427.561194852" Oct 06 10:43:00 crc kubenswrapper[4989]: I1006 10:43:00.230407 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:00 crc kubenswrapper[4989]: I1006 10:43:00.231242 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:00 crc kubenswrapper[4989]: I1006 10:43:00.306382 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:03 crc kubenswrapper[4989]: I1006 10:43:03.938894 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:43:03 crc kubenswrapper[4989]: I1006 10:43:03.939453 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:43:10 crc kubenswrapper[4989]: I1006 10:43:10.292393 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:10 crc kubenswrapper[4989]: I1006 10:43:10.340184 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:43:10 crc kubenswrapper[4989]: I1006 10:43:10.884382 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dd2m7" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="registry-server" containerID="cri-o://7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5" gracePeriod=2 Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.429441 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.613219 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-952lr\" (UniqueName: \"kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr\") pod \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.613344 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content\") pod \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.613434 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities\") pod \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\" (UID: \"9e5d109b-5aea-482a-9bef-5a8ef525c45f\") " Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.614873 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities" (OuterVolumeSpecName: "utilities") pod "9e5d109b-5aea-482a-9bef-5a8ef525c45f" (UID: "9e5d109b-5aea-482a-9bef-5a8ef525c45f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.621311 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr" (OuterVolumeSpecName: "kube-api-access-952lr") pod "9e5d109b-5aea-482a-9bef-5a8ef525c45f" (UID: "9e5d109b-5aea-482a-9bef-5a8ef525c45f"). InnerVolumeSpecName "kube-api-access-952lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.660492 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e5d109b-5aea-482a-9bef-5a8ef525c45f" (UID: "9e5d109b-5aea-482a-9bef-5a8ef525c45f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.716952 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-952lr\" (UniqueName: \"kubernetes.io/projected/9e5d109b-5aea-482a-9bef-5a8ef525c45f-kube-api-access-952lr\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.717004 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.717022 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e5d109b-5aea-482a-9bef-5a8ef525c45f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.894700 4989 generic.go:334] "Generic (PLEG): container finished" podID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerID="7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5" exitCode=0 Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.894749 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerDied","Data":"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5"} Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.894794 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dd2m7" event={"ID":"9e5d109b-5aea-482a-9bef-5a8ef525c45f","Type":"ContainerDied","Data":"945cbc97b61ee138c83e796cc983fd55e5b7982c4566ae69f70c47ea2bacd2c5"} Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.894816 4989 scope.go:117] "RemoveContainer" containerID="7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.894832 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dd2m7" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.944666 4989 scope.go:117] "RemoveContainer" containerID="d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a" Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.953788 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.953845 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dd2m7"] Oct 06 10:43:11 crc kubenswrapper[4989]: I1006 10:43:11.965950 4989 scope.go:117] "RemoveContainer" containerID="b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.013817 4989 scope.go:117] "RemoveContainer" containerID="7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5" Oct 06 10:43:12 crc kubenswrapper[4989]: E1006 10:43:12.014510 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5\": container with ID starting with 7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5 not found: ID does not exist" containerID="7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.014549 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5"} err="failed to get container status \"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5\": rpc error: code = NotFound desc = could not find container \"7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5\": container with ID starting with 7d3b78869efd8026b7292983f261849c005809e84884d72814115b94557a80e5 not found: ID does not exist" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.014573 4989 scope.go:117] "RemoveContainer" containerID="d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a" Oct 06 10:43:12 crc kubenswrapper[4989]: E1006 10:43:12.015095 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a\": container with ID starting with d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a not found: ID does not exist" containerID="d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.015123 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a"} err="failed to get container status \"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a\": rpc error: code = NotFound desc = could not find container \"d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a\": container with ID starting with d3062e893972eb374d88b9e16474a4f268bf314919c42b5c6299419dde11f30a not found: ID does not exist" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.015140 4989 scope.go:117] "RemoveContainer" containerID="b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462" Oct 06 10:43:12 crc kubenswrapper[4989]: E1006 10:43:12.015434 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462\": container with ID starting with b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462 not found: ID does not exist" containerID="b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462" Oct 06 10:43:12 crc kubenswrapper[4989]: I1006 10:43:12.015463 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462"} err="failed to get container status \"b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462\": rpc error: code = NotFound desc = could not find container \"b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462\": container with ID starting with b67d06476961e47ad8ce613926cbb77a7ce7c99801293e342e1224344af6e462 not found: ID does not exist" Oct 06 10:43:13 crc kubenswrapper[4989]: I1006 10:43:13.950136 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" path="/var/lib/kubelet/pods/9e5d109b-5aea-482a-9bef-5a8ef525c45f/volumes" Oct 06 10:43:33 crc kubenswrapper[4989]: I1006 10:43:33.936149 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:43:33 crc kubenswrapper[4989]: I1006 10:43:33.936984 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:43:33 crc kubenswrapper[4989]: I1006 10:43:33.954120 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:43:33 crc kubenswrapper[4989]: I1006 10:43:33.955279 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:43:33 crc kubenswrapper[4989]: I1006 10:43:33.955377 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" gracePeriod=600 Oct 06 10:43:34 crc kubenswrapper[4989]: I1006 10:43:34.131344 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" exitCode=0 Oct 06 10:43:34 crc kubenswrapper[4989]: I1006 10:43:34.131395 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e"} Oct 06 10:43:34 crc kubenswrapper[4989]: I1006 10:43:34.131433 4989 scope.go:117] "RemoveContainer" containerID="a142ff863d43e96357139e2d1695653959252267af205b6d5d70864a10f4dff7" Oct 06 10:43:34 crc kubenswrapper[4989]: E1006 10:43:34.228026 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:43:35 crc kubenswrapper[4989]: I1006 10:43:35.150204 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:43:35 crc kubenswrapper[4989]: E1006 10:43:35.153741 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:43:49 crc kubenswrapper[4989]: I1006 10:43:49.949419 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:43:49 crc kubenswrapper[4989]: E1006 10:43:49.950302 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:43:55 crc kubenswrapper[4989]: I1006 10:43:55.393627 4989 generic.go:334] "Generic (PLEG): container finished" podID="8f67e476-79a2-499d-ad1b-142eac5e3681" containerID="8c99b6a24af8675892eb960657745f9560dd91f176390eeb74c373c59582b7f8" exitCode=0 Oct 06 10:43:55 crc kubenswrapper[4989]: I1006 10:43:55.393729 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-65277" event={"ID":"8f67e476-79a2-499d-ad1b-142eac5e3681","Type":"ContainerDied","Data":"8c99b6a24af8675892eb960657745f9560dd91f176390eeb74c373c59582b7f8"} Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.891591 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.923521 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key\") pod \"8f67e476-79a2-499d-ad1b-142eac5e3681\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.925332 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory\") pod \"8f67e476-79a2-499d-ad1b-142eac5e3681\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.925433 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph\") pod \"8f67e476-79a2-499d-ad1b-142eac5e3681\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.925773 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmrn\" (UniqueName: \"kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn\") pod \"8f67e476-79a2-499d-ad1b-142eac5e3681\" (UID: \"8f67e476-79a2-499d-ad1b-142eac5e3681\") " Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.930526 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn" (OuterVolumeSpecName: "kube-api-access-zlmrn") pod "8f67e476-79a2-499d-ad1b-142eac5e3681" (UID: "8f67e476-79a2-499d-ad1b-142eac5e3681"). InnerVolumeSpecName "kube-api-access-zlmrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.930848 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph" (OuterVolumeSpecName: "ceph") pod "8f67e476-79a2-499d-ad1b-142eac5e3681" (UID: "8f67e476-79a2-499d-ad1b-142eac5e3681"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.956871 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8f67e476-79a2-499d-ad1b-142eac5e3681" (UID: "8f67e476-79a2-499d-ad1b-142eac5e3681"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:43:56 crc kubenswrapper[4989]: I1006 10:43:56.966347 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory" (OuterVolumeSpecName: "inventory") pod "8f67e476-79a2-499d-ad1b-142eac5e3681" (UID: "8f67e476-79a2-499d-ad1b-142eac5e3681"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.031152 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.031195 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.031209 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmrn\" (UniqueName: \"kubernetes.io/projected/8f67e476-79a2-499d-ad1b-142eac5e3681-kube-api-access-zlmrn\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.031222 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8f67e476-79a2-499d-ad1b-142eac5e3681-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.416082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-65277" event={"ID":"8f67e476-79a2-499d-ad1b-142eac5e3681","Type":"ContainerDied","Data":"160f0be5bae66876732d4f0665e3c7f3f6cfe8e9fc161cea1bf9c1442bcdf049"} Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.416428 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="160f0be5bae66876732d4f0665e3c7f3f6cfe8e9fc161cea1bf9c1442bcdf049" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.416121 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-65277" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.533877 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9c25w"] Oct 06 10:43:57 crc kubenswrapper[4989]: E1006 10:43:57.534832 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="extract-content" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.534937 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="extract-content" Oct 06 10:43:57 crc kubenswrapper[4989]: E1006 10:43:57.535026 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f67e476-79a2-499d-ad1b-142eac5e3681" containerName="download-cache-openstack-openstack-cell1" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.535112 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f67e476-79a2-499d-ad1b-142eac5e3681" containerName="download-cache-openstack-openstack-cell1" Oct 06 10:43:57 crc kubenswrapper[4989]: E1006 10:43:57.535206 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="extract-utilities" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.535274 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="extract-utilities" Oct 06 10:43:57 crc kubenswrapper[4989]: E1006 10:43:57.535365 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="registry-server" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.535438 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="registry-server" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.535842 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5d109b-5aea-482a-9bef-5a8ef525c45f" containerName="registry-server" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.535957 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f67e476-79a2-499d-ad1b-142eac5e3681" containerName="download-cache-openstack-openstack-cell1" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.537032 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.552326 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.552635 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.553272 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.553862 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.598259 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9c25w"] Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.645104 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.645244 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.645321 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dxm6\" (UniqueName: \"kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.645353 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.748047 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.748191 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dxm6\" (UniqueName: \"kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.748236 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.748418 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.754992 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.761689 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.764819 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.771503 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dxm6\" (UniqueName: \"kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6\") pod \"configure-network-openstack-openstack-cell1-9c25w\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:57 crc kubenswrapper[4989]: I1006 10:43:57.886800 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:43:58 crc kubenswrapper[4989]: I1006 10:43:58.437814 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-9c25w"] Oct 06 10:43:59 crc kubenswrapper[4989]: I1006 10:43:59.442470 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" event={"ID":"bf37b711-adc9-4b80-b9fc-efc5dcced79a","Type":"ContainerStarted","Data":"9b62c69522131a4a728fb13bb44c93510d59b308cc0670f5a9d1caf7241a669d"} Oct 06 10:43:59 crc kubenswrapper[4989]: I1006 10:43:59.442844 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" event={"ID":"bf37b711-adc9-4b80-b9fc-efc5dcced79a","Type":"ContainerStarted","Data":"9ebdcc9070c3715d3ad16e46ab6f461b37b91208ea40d898ef2e04e7c2243868"} Oct 06 10:43:59 crc kubenswrapper[4989]: I1006 10:43:59.470081 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" podStartSLOduration=2.289146668 podStartE2EDuration="2.470064095s" podCreationTimestamp="2025-10-06 10:43:57 +0000 UTC" firstStartedPulling="2025-10-06 10:43:58.449475872 +0000 UTC m=+7489.239501452" lastFinishedPulling="2025-10-06 10:43:58.630393299 +0000 UTC m=+7489.420418879" observedRunningTime="2025-10-06 10:43:59.463316742 +0000 UTC m=+7490.253342332" watchObservedRunningTime="2025-10-06 10:43:59.470064095 +0000 UTC m=+7490.260089675" Oct 06 10:44:03 crc kubenswrapper[4989]: I1006 10:44:03.936381 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:44:03 crc kubenswrapper[4989]: E1006 10:44:03.937196 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:44:18 crc kubenswrapper[4989]: I1006 10:44:18.937014 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:44:18 crc kubenswrapper[4989]: E1006 10:44:18.938117 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:44:32 crc kubenswrapper[4989]: I1006 10:44:32.935962 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:44:32 crc kubenswrapper[4989]: E1006 10:44:32.936752 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:44:45 crc kubenswrapper[4989]: I1006 10:44:45.936468 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:44:45 crc kubenswrapper[4989]: E1006 10:44:45.937735 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:44:56 crc kubenswrapper[4989]: I1006 10:44:56.935462 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:44:56 crc kubenswrapper[4989]: E1006 10:44:56.936274 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.176133 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw"] Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.180468 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.182905 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.183007 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.210877 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw"] Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.303212 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hbsp\" (UniqueName: \"kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.303517 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.303570 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.405482 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.405550 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.405685 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hbsp\" (UniqueName: \"kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.406579 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.412145 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.430971 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hbsp\" (UniqueName: \"kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp\") pod \"collect-profiles-29329125-kbtmw\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.511584 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:00 crc kubenswrapper[4989]: I1006 10:45:00.994209 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw"] Oct 06 10:45:01 crc kubenswrapper[4989]: I1006 10:45:01.098507 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" event={"ID":"3bc87c85-b9ee-48b5-a357-015a136c798e","Type":"ContainerStarted","Data":"8956ea63fbeac9f64dd9bc014156feaf2267e9912f0dbe18eef72f35b8a1d65a"} Oct 06 10:45:02 crc kubenswrapper[4989]: I1006 10:45:02.109682 4989 generic.go:334] "Generic (PLEG): container finished" podID="3bc87c85-b9ee-48b5-a357-015a136c798e" containerID="4e91ae88869d41b812769fbd27c3c30a418ac5e421058a59d886552807444638" exitCode=0 Oct 06 10:45:02 crc kubenswrapper[4989]: I1006 10:45:02.109792 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" event={"ID":"3bc87c85-b9ee-48b5-a357-015a136c798e","Type":"ContainerDied","Data":"4e91ae88869d41b812769fbd27c3c30a418ac5e421058a59d886552807444638"} Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.509781 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.677987 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume\") pod \"3bc87c85-b9ee-48b5-a357-015a136c798e\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.678453 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume\") pod \"3bc87c85-b9ee-48b5-a357-015a136c798e\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.678535 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hbsp\" (UniqueName: \"kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp\") pod \"3bc87c85-b9ee-48b5-a357-015a136c798e\" (UID: \"3bc87c85-b9ee-48b5-a357-015a136c798e\") " Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.679097 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3bc87c85-b9ee-48b5-a357-015a136c798e" (UID: "3bc87c85-b9ee-48b5-a357-015a136c798e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.679271 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3bc87c85-b9ee-48b5-a357-015a136c798e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.684595 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp" (OuterVolumeSpecName: "kube-api-access-4hbsp") pod "3bc87c85-b9ee-48b5-a357-015a136c798e" (UID: "3bc87c85-b9ee-48b5-a357-015a136c798e"). InnerVolumeSpecName "kube-api-access-4hbsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.684609 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3bc87c85-b9ee-48b5-a357-015a136c798e" (UID: "3bc87c85-b9ee-48b5-a357-015a136c798e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.780849 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hbsp\" (UniqueName: \"kubernetes.io/projected/3bc87c85-b9ee-48b5-a357-015a136c798e-kube-api-access-4hbsp\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:03 crc kubenswrapper[4989]: I1006 10:45:03.780886 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3bc87c85-b9ee-48b5-a357-015a136c798e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:04 crc kubenswrapper[4989]: I1006 10:45:04.130467 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" event={"ID":"3bc87c85-b9ee-48b5-a357-015a136c798e","Type":"ContainerDied","Data":"8956ea63fbeac9f64dd9bc014156feaf2267e9912f0dbe18eef72f35b8a1d65a"} Oct 06 10:45:04 crc kubenswrapper[4989]: I1006 10:45:04.130833 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8956ea63fbeac9f64dd9bc014156feaf2267e9912f0dbe18eef72f35b8a1d65a" Oct 06 10:45:04 crc kubenswrapper[4989]: I1006 10:45:04.130792 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw" Oct 06 10:45:04 crc kubenswrapper[4989]: I1006 10:45:04.599417 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94"] Oct 06 10:45:04 crc kubenswrapper[4989]: I1006 10:45:04.611635 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329080-g4s94"] Oct 06 10:45:05 crc kubenswrapper[4989]: I1006 10:45:05.949646 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66463be1-c057-460f-96a2-1ed922b8621d" path="/var/lib/kubelet/pods/66463be1-c057-460f-96a2-1ed922b8621d/volumes" Oct 06 10:45:08 crc kubenswrapper[4989]: I1006 10:45:08.936652 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:45:08 crc kubenswrapper[4989]: E1006 10:45:08.937273 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:45:12 crc kubenswrapper[4989]: E1006 10:45:12.761436 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf37b711_adc9_4b80_b9fc_efc5dcced79a.slice/crio-9b62c69522131a4a728fb13bb44c93510d59b308cc0670f5a9d1caf7241a669d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf37b711_adc9_4b80_b9fc_efc5dcced79a.slice/crio-conmon-9b62c69522131a4a728fb13bb44c93510d59b308cc0670f5a9d1caf7241a669d.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:45:13 crc kubenswrapper[4989]: I1006 10:45:13.215579 4989 generic.go:334] "Generic (PLEG): container finished" podID="bf37b711-adc9-4b80-b9fc-efc5dcced79a" containerID="9b62c69522131a4a728fb13bb44c93510d59b308cc0670f5a9d1caf7241a669d" exitCode=0 Oct 06 10:45:13 crc kubenswrapper[4989]: I1006 10:45:13.215718 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" event={"ID":"bf37b711-adc9-4b80-b9fc-efc5dcced79a","Type":"ContainerDied","Data":"9b62c69522131a4a728fb13bb44c93510d59b308cc0670f5a9d1caf7241a669d"} Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.704977 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.818163 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dxm6\" (UniqueName: \"kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6\") pod \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.818322 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory\") pod \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.818414 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph\") pod \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.818459 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key\") pod \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\" (UID: \"bf37b711-adc9-4b80-b9fc-efc5dcced79a\") " Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.823717 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph" (OuterVolumeSpecName: "ceph") pod "bf37b711-adc9-4b80-b9fc-efc5dcced79a" (UID: "bf37b711-adc9-4b80-b9fc-efc5dcced79a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.824485 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6" (OuterVolumeSpecName: "kube-api-access-2dxm6") pod "bf37b711-adc9-4b80-b9fc-efc5dcced79a" (UID: "bf37b711-adc9-4b80-b9fc-efc5dcced79a"). InnerVolumeSpecName "kube-api-access-2dxm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.847596 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf37b711-adc9-4b80-b9fc-efc5dcced79a" (UID: "bf37b711-adc9-4b80-b9fc-efc5dcced79a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.847924 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory" (OuterVolumeSpecName: "inventory") pod "bf37b711-adc9-4b80-b9fc-efc5dcced79a" (UID: "bf37b711-adc9-4b80-b9fc-efc5dcced79a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.921579 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.921958 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.921971 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf37b711-adc9-4b80-b9fc-efc5dcced79a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:14 crc kubenswrapper[4989]: I1006 10:45:14.921984 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dxm6\" (UniqueName: \"kubernetes.io/projected/bf37b711-adc9-4b80-b9fc-efc5dcced79a-kube-api-access-2dxm6\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.239495 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" event={"ID":"bf37b711-adc9-4b80-b9fc-efc5dcced79a","Type":"ContainerDied","Data":"9ebdcc9070c3715d3ad16e46ab6f461b37b91208ea40d898ef2e04e7c2243868"} Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.239535 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-9c25w" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.239543 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ebdcc9070c3715d3ad16e46ab6f461b37b91208ea40d898ef2e04e7c2243868" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.318549 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-5cjdc"] Oct 06 10:45:15 crc kubenswrapper[4989]: E1006 10:45:15.319121 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf37b711-adc9-4b80-b9fc-efc5dcced79a" containerName="configure-network-openstack-openstack-cell1" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.319143 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf37b711-adc9-4b80-b9fc-efc5dcced79a" containerName="configure-network-openstack-openstack-cell1" Oct 06 10:45:15 crc kubenswrapper[4989]: E1006 10:45:15.319201 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc87c85-b9ee-48b5-a357-015a136c798e" containerName="collect-profiles" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.319211 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc87c85-b9ee-48b5-a357-015a136c798e" containerName="collect-profiles" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.319460 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc87c85-b9ee-48b5-a357-015a136c798e" containerName="collect-profiles" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.319492 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf37b711-adc9-4b80-b9fc-efc5dcced79a" containerName="configure-network-openstack-openstack-cell1" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.320400 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.328214 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-5cjdc"] Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.339922 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.340140 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.340304 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.340534 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.431799 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.431904 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.432063 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.432101 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.534533 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.534634 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.534764 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.534793 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.538972 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.539711 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.540367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.553724 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g\") pod \"validate-network-openstack-openstack-cell1-5cjdc\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:15 crc kubenswrapper[4989]: I1006 10:45:15.654308 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:16 crc kubenswrapper[4989]: I1006 10:45:16.208615 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-5cjdc"] Oct 06 10:45:16 crc kubenswrapper[4989]: I1006 10:45:16.252533 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" event={"ID":"338161e0-27f7-4a63-86c8-1e35e92a352c","Type":"ContainerStarted","Data":"e61f26cb3b6dfe47836c6a792980bc7de7f9b8e2cad97ba9918c055464a2ae63"} Oct 06 10:45:17 crc kubenswrapper[4989]: I1006 10:45:17.264023 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" event={"ID":"338161e0-27f7-4a63-86c8-1e35e92a352c","Type":"ContainerStarted","Data":"277907c9f741162fb7a92847a7f74d9bebd2d84478c04c414dd8fa11c664be0c"} Oct 06 10:45:17 crc kubenswrapper[4989]: I1006 10:45:17.291809 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" podStartSLOduration=2.116168549 podStartE2EDuration="2.291772123s" podCreationTimestamp="2025-10-06 10:45:15 +0000 UTC" firstStartedPulling="2025-10-06 10:45:16.217907045 +0000 UTC m=+7567.007932635" lastFinishedPulling="2025-10-06 10:45:16.393510629 +0000 UTC m=+7567.183536209" observedRunningTime="2025-10-06 10:45:17.279004558 +0000 UTC m=+7568.069030168" watchObservedRunningTime="2025-10-06 10:45:17.291772123 +0000 UTC m=+7568.081797713" Oct 06 10:45:22 crc kubenswrapper[4989]: I1006 10:45:22.326174 4989 generic.go:334] "Generic (PLEG): container finished" podID="338161e0-27f7-4a63-86c8-1e35e92a352c" containerID="277907c9f741162fb7a92847a7f74d9bebd2d84478c04c414dd8fa11c664be0c" exitCode=0 Oct 06 10:45:22 crc kubenswrapper[4989]: I1006 10:45:22.326248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" event={"ID":"338161e0-27f7-4a63-86c8-1e35e92a352c","Type":"ContainerDied","Data":"277907c9f741162fb7a92847a7f74d9bebd2d84478c04c414dd8fa11c664be0c"} Oct 06 10:45:23 crc kubenswrapper[4989]: I1006 10:45:23.896691 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:23 crc kubenswrapper[4989]: I1006 10:45:23.936056 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:45:23 crc kubenswrapper[4989]: E1006 10:45:23.936365 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.043488 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g\") pod \"338161e0-27f7-4a63-86c8-1e35e92a352c\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.043559 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph\") pod \"338161e0-27f7-4a63-86c8-1e35e92a352c\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.043631 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory\") pod \"338161e0-27f7-4a63-86c8-1e35e92a352c\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.043676 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key\") pod \"338161e0-27f7-4a63-86c8-1e35e92a352c\" (UID: \"338161e0-27f7-4a63-86c8-1e35e92a352c\") " Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.051002 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g" (OuterVolumeSpecName: "kube-api-access-xcv8g") pod "338161e0-27f7-4a63-86c8-1e35e92a352c" (UID: "338161e0-27f7-4a63-86c8-1e35e92a352c"). InnerVolumeSpecName "kube-api-access-xcv8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.051982 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph" (OuterVolumeSpecName: "ceph") pod "338161e0-27f7-4a63-86c8-1e35e92a352c" (UID: "338161e0-27f7-4a63-86c8-1e35e92a352c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.081157 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "338161e0-27f7-4a63-86c8-1e35e92a352c" (UID: "338161e0-27f7-4a63-86c8-1e35e92a352c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.083009 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory" (OuterVolumeSpecName: "inventory") pod "338161e0-27f7-4a63-86c8-1e35e92a352c" (UID: "338161e0-27f7-4a63-86c8-1e35e92a352c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.146708 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.146795 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcv8g\" (UniqueName: \"kubernetes.io/projected/338161e0-27f7-4a63-86c8-1e35e92a352c-kube-api-access-xcv8g\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.146816 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.146833 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/338161e0-27f7-4a63-86c8-1e35e92a352c-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.353694 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.353698 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-5cjdc" event={"ID":"338161e0-27f7-4a63-86c8-1e35e92a352c","Type":"ContainerDied","Data":"e61f26cb3b6dfe47836c6a792980bc7de7f9b8e2cad97ba9918c055464a2ae63"} Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.353884 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e61f26cb3b6dfe47836c6a792980bc7de7f9b8e2cad97ba9918c055464a2ae63" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.437431 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-gwb5g"] Oct 06 10:45:24 crc kubenswrapper[4989]: E1006 10:45:24.438049 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338161e0-27f7-4a63-86c8-1e35e92a352c" containerName="validate-network-openstack-openstack-cell1" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.438070 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="338161e0-27f7-4a63-86c8-1e35e92a352c" containerName="validate-network-openstack-openstack-cell1" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.438300 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="338161e0-27f7-4a63-86c8-1e35e92a352c" containerName="validate-network-openstack-openstack-cell1" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.439117 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.445220 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-gwb5g"] Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.445549 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.446020 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.446149 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.446256 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.555068 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.555384 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6czf\" (UniqueName: \"kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.555496 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.555639 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.659017 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6czf\" (UniqueName: \"kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.661787 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.661822 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.662236 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.667816 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.670358 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.686020 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.691110 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6czf\" (UniqueName: \"kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf\") pod \"install-os-openstack-openstack-cell1-gwb5g\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:24 crc kubenswrapper[4989]: I1006 10:45:24.768287 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:45:25 crc kubenswrapper[4989]: I1006 10:45:25.336349 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-gwb5g"] Oct 06 10:45:25 crc kubenswrapper[4989]: I1006 10:45:25.349192 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:45:25 crc kubenswrapper[4989]: I1006 10:45:25.363606 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" event={"ID":"5204bd93-cd20-44b3-9f50-8f4113d8ab62","Type":"ContainerStarted","Data":"eb88c65e5dadbf01dc01b7c19579788058de84a127caa979942549e6fba2b9e8"} Oct 06 10:45:26 crc kubenswrapper[4989]: I1006 10:45:26.375559 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" event={"ID":"5204bd93-cd20-44b3-9f50-8f4113d8ab62","Type":"ContainerStarted","Data":"4f26113d72835dddeeab5999830e6a96ec7c5f2c3c761e2e98c1f3db5b5738c7"} Oct 06 10:45:26 crc kubenswrapper[4989]: I1006 10:45:26.401284 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" podStartSLOduration=2.229272458 podStartE2EDuration="2.401261839s" podCreationTimestamp="2025-10-06 10:45:24 +0000 UTC" firstStartedPulling="2025-10-06 10:45:25.348973818 +0000 UTC m=+7576.138999398" lastFinishedPulling="2025-10-06 10:45:25.520963199 +0000 UTC m=+7576.310988779" observedRunningTime="2025-10-06 10:45:26.39571199 +0000 UTC m=+7577.185737580" watchObservedRunningTime="2025-10-06 10:45:26.401261839 +0000 UTC m=+7577.191287419" Oct 06 10:45:27 crc kubenswrapper[4989]: I1006 10:45:27.122092 4989 scope.go:117] "RemoveContainer" containerID="676064bb353c3f0ca4e8ea4d0a568248559ef9d5fdb650ca5259948737657dd9" Oct 06 10:45:36 crc kubenswrapper[4989]: I1006 10:45:36.936680 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:45:36 crc kubenswrapper[4989]: E1006 10:45:36.938484 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.594878 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.600065 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.606415 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.690077 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dksb\" (UniqueName: \"kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.690233 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.690380 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.792044 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.792179 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dksb\" (UniqueName: \"kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.792306 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.792487 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.793119 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.818736 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dksb\" (UniqueName: \"kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb\") pod \"redhat-operators-g9gsv\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:43 crc kubenswrapper[4989]: I1006 10:45:43.938629 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:44 crc kubenswrapper[4989]: I1006 10:45:44.422725 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:44 crc kubenswrapper[4989]: I1006 10:45:44.556443 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerStarted","Data":"cba2c6d7d39925cf1a819fea3f95620e457712532b1e1c18dbdcdfca255c48ea"} Oct 06 10:45:45 crc kubenswrapper[4989]: I1006 10:45:45.569801 4989 generic.go:334] "Generic (PLEG): container finished" podID="104df711-83db-4e2e-a47b-b311e212c1f0" containerID="a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a" exitCode=0 Oct 06 10:45:45 crc kubenswrapper[4989]: I1006 10:45:45.570188 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerDied","Data":"a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a"} Oct 06 10:45:47 crc kubenswrapper[4989]: I1006 10:45:47.591592 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerStarted","Data":"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8"} Oct 06 10:45:48 crc kubenswrapper[4989]: I1006 10:45:48.602128 4989 generic.go:334] "Generic (PLEG): container finished" podID="104df711-83db-4e2e-a47b-b311e212c1f0" containerID="a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8" exitCode=0 Oct 06 10:45:48 crc kubenswrapper[4989]: I1006 10:45:48.602202 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerDied","Data":"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8"} Oct 06 10:45:49 crc kubenswrapper[4989]: I1006 10:45:49.614226 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerStarted","Data":"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5"} Oct 06 10:45:49 crc kubenswrapper[4989]: I1006 10:45:49.943693 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:45:49 crc kubenswrapper[4989]: E1006 10:45:49.943933 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:45:53 crc kubenswrapper[4989]: I1006 10:45:53.960188 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:53 crc kubenswrapper[4989]: I1006 10:45:53.961892 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:53 crc kubenswrapper[4989]: I1006 10:45:53.998761 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:54 crc kubenswrapper[4989]: I1006 10:45:54.039158 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g9gsv" podStartSLOduration=7.487007746 podStartE2EDuration="11.03913748s" podCreationTimestamp="2025-10-06 10:45:43 +0000 UTC" firstStartedPulling="2025-10-06 10:45:45.572564801 +0000 UTC m=+7596.362590411" lastFinishedPulling="2025-10-06 10:45:49.124694565 +0000 UTC m=+7599.914720145" observedRunningTime="2025-10-06 10:45:49.629782738 +0000 UTC m=+7600.419808318" watchObservedRunningTime="2025-10-06 10:45:54.03913748 +0000 UTC m=+7604.829163080" Oct 06 10:45:54 crc kubenswrapper[4989]: I1006 10:45:54.724971 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:54 crc kubenswrapper[4989]: I1006 10:45:54.786925 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:56 crc kubenswrapper[4989]: I1006 10:45:56.686829 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g9gsv" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="registry-server" containerID="cri-o://c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5" gracePeriod=2 Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.174729 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.195334 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dksb\" (UniqueName: \"kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb\") pod \"104df711-83db-4e2e-a47b-b311e212c1f0\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.195625 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities\") pod \"104df711-83db-4e2e-a47b-b311e212c1f0\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.195902 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content\") pod \"104df711-83db-4e2e-a47b-b311e212c1f0\" (UID: \"104df711-83db-4e2e-a47b-b311e212c1f0\") " Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.197873 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities" (OuterVolumeSpecName: "utilities") pod "104df711-83db-4e2e-a47b-b311e212c1f0" (UID: "104df711-83db-4e2e-a47b-b311e212c1f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.198616 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.240730 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb" (OuterVolumeSpecName: "kube-api-access-4dksb") pod "104df711-83db-4e2e-a47b-b311e212c1f0" (UID: "104df711-83db-4e2e-a47b-b311e212c1f0"). InnerVolumeSpecName "kube-api-access-4dksb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.297841 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "104df711-83db-4e2e-a47b-b311e212c1f0" (UID: "104df711-83db-4e2e-a47b-b311e212c1f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.300158 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dksb\" (UniqueName: \"kubernetes.io/projected/104df711-83db-4e2e-a47b-b311e212c1f0-kube-api-access-4dksb\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.300208 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/104df711-83db-4e2e-a47b-b311e212c1f0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.696679 4989 generic.go:334] "Generic (PLEG): container finished" podID="104df711-83db-4e2e-a47b-b311e212c1f0" containerID="c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5" exitCode=0 Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.696723 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerDied","Data":"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5"} Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.696754 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g9gsv" event={"ID":"104df711-83db-4e2e-a47b-b311e212c1f0","Type":"ContainerDied","Data":"cba2c6d7d39925cf1a819fea3f95620e457712532b1e1c18dbdcdfca255c48ea"} Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.696773 4989 scope.go:117] "RemoveContainer" containerID="c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.696915 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g9gsv" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.735771 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.744856 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g9gsv"] Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.778585 4989 scope.go:117] "RemoveContainer" containerID="a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.802716 4989 scope.go:117] "RemoveContainer" containerID="a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.844555 4989 scope.go:117] "RemoveContainer" containerID="c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5" Oct 06 10:45:57 crc kubenswrapper[4989]: E1006 10:45:57.844995 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5\": container with ID starting with c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5 not found: ID does not exist" containerID="c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.845034 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5"} err="failed to get container status \"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5\": rpc error: code = NotFound desc = could not find container \"c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5\": container with ID starting with c0a99f80d93f15c2fc51affab3c45fa9998d832b141e45501442aba2890238c5 not found: ID does not exist" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.845080 4989 scope.go:117] "RemoveContainer" containerID="a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8" Oct 06 10:45:57 crc kubenswrapper[4989]: E1006 10:45:57.845440 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8\": container with ID starting with a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8 not found: ID does not exist" containerID="a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.845473 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8"} err="failed to get container status \"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8\": rpc error: code = NotFound desc = could not find container \"a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8\": container with ID starting with a8a5340031ff066990f98bad04068ca9b73c6329610ee719f4b46f85177f7ea8 not found: ID does not exist" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.845487 4989 scope.go:117] "RemoveContainer" containerID="a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a" Oct 06 10:45:57 crc kubenswrapper[4989]: E1006 10:45:57.845707 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a\": container with ID starting with a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a not found: ID does not exist" containerID="a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.845733 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a"} err="failed to get container status \"a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a\": rpc error: code = NotFound desc = could not find container \"a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a\": container with ID starting with a7a919ce149452aab580d31bada448084f86da71aad5593b8afc78b37f3f854a not found: ID does not exist" Oct 06 10:45:57 crc kubenswrapper[4989]: I1006 10:45:57.953246 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" path="/var/lib/kubelet/pods/104df711-83db-4e2e-a47b-b311e212c1f0/volumes" Oct 06 10:46:02 crc kubenswrapper[4989]: I1006 10:46:02.937577 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:46:02 crc kubenswrapper[4989]: E1006 10:46:02.938915 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:46:10 crc kubenswrapper[4989]: I1006 10:46:10.850462 4989 generic.go:334] "Generic (PLEG): container finished" podID="5204bd93-cd20-44b3-9f50-8f4113d8ab62" containerID="4f26113d72835dddeeab5999830e6a96ec7c5f2c3c761e2e98c1f3db5b5738c7" exitCode=0 Oct 06 10:46:10 crc kubenswrapper[4989]: I1006 10:46:10.850595 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" event={"ID":"5204bd93-cd20-44b3-9f50-8f4113d8ab62","Type":"ContainerDied","Data":"4f26113d72835dddeeab5999830e6a96ec7c5f2c3c761e2e98c1f3db5b5738c7"} Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.334418 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.446993 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory\") pod \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.447194 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6czf\" (UniqueName: \"kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf\") pod \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.447304 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key\") pod \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.447366 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph\") pod \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\" (UID: \"5204bd93-cd20-44b3-9f50-8f4113d8ab62\") " Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.451845 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph" (OuterVolumeSpecName: "ceph") pod "5204bd93-cd20-44b3-9f50-8f4113d8ab62" (UID: "5204bd93-cd20-44b3-9f50-8f4113d8ab62"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.452899 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf" (OuterVolumeSpecName: "kube-api-access-m6czf") pod "5204bd93-cd20-44b3-9f50-8f4113d8ab62" (UID: "5204bd93-cd20-44b3-9f50-8f4113d8ab62"). InnerVolumeSpecName "kube-api-access-m6czf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.482117 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5204bd93-cd20-44b3-9f50-8f4113d8ab62" (UID: "5204bd93-cd20-44b3-9f50-8f4113d8ab62"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.505537 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory" (OuterVolumeSpecName: "inventory") pod "5204bd93-cd20-44b3-9f50-8f4113d8ab62" (UID: "5204bd93-cd20-44b3-9f50-8f4113d8ab62"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.550367 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.550396 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.550405 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5204bd93-cd20-44b3-9f50-8f4113d8ab62-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.550414 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6czf\" (UniqueName: \"kubernetes.io/projected/5204bd93-cd20-44b3-9f50-8f4113d8ab62-kube-api-access-m6czf\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.882776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" event={"ID":"5204bd93-cd20-44b3-9f50-8f4113d8ab62","Type":"ContainerDied","Data":"eb88c65e5dadbf01dc01b7c19579788058de84a127caa979942549e6fba2b9e8"} Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.882838 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb88c65e5dadbf01dc01b7c19579788058de84a127caa979942549e6fba2b9e8" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.882889 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-gwb5g" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.985605 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2xlln"] Oct 06 10:46:12 crc kubenswrapper[4989]: E1006 10:46:12.986760 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="extract-utilities" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.986785 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="extract-utilities" Oct 06 10:46:12 crc kubenswrapper[4989]: E1006 10:46:12.986832 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="extract-content" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.986841 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="extract-content" Oct 06 10:46:12 crc kubenswrapper[4989]: E1006 10:46:12.986860 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="registry-server" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.986868 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="registry-server" Oct 06 10:46:12 crc kubenswrapper[4989]: E1006 10:46:12.986896 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5204bd93-cd20-44b3-9f50-8f4113d8ab62" containerName="install-os-openstack-openstack-cell1" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.986904 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5204bd93-cd20-44b3-9f50-8f4113d8ab62" containerName="install-os-openstack-openstack-cell1" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.987160 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="104df711-83db-4e2e-a47b-b311e212c1f0" containerName="registry-server" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.987196 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5204bd93-cd20-44b3-9f50-8f4113d8ab62" containerName="install-os-openstack-openstack-cell1" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.988081 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.990517 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.991115 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.991464 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:46:12 crc kubenswrapper[4989]: I1006 10:46:12.992384 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.011530 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2xlln"] Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.164310 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl8gn\" (UniqueName: \"kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.164430 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.164538 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.164648 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.267207 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl8gn\" (UniqueName: \"kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.267378 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.267483 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.267530 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.271392 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.272953 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.273138 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.289232 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl8gn\" (UniqueName: \"kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn\") pod \"configure-os-openstack-openstack-cell1-2xlln\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.318463 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.876883 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-2xlln"] Oct 06 10:46:13 crc kubenswrapper[4989]: I1006 10:46:13.900066 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" event={"ID":"ee637d80-17d5-4a6c-9734-67f46cdc47f0","Type":"ContainerStarted","Data":"d15c79ad0c680065fb32d66ec99d753df6020d01a2a614a41b212dfd02711313"} Oct 06 10:46:14 crc kubenswrapper[4989]: I1006 10:46:14.911960 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" event={"ID":"ee637d80-17d5-4a6c-9734-67f46cdc47f0","Type":"ContainerStarted","Data":"3bf2f250819332d82fea007ccbbd2d1cfbde62ecf65edb6c188fe914a824150c"} Oct 06 10:46:14 crc kubenswrapper[4989]: I1006 10:46:14.932148 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" podStartSLOduration=2.736747616 podStartE2EDuration="2.932133677s" podCreationTimestamp="2025-10-06 10:46:12 +0000 UTC" firstStartedPulling="2025-10-06 10:46:13.889946176 +0000 UTC m=+7624.679971766" lastFinishedPulling="2025-10-06 10:46:14.085332247 +0000 UTC m=+7624.875357827" observedRunningTime="2025-10-06 10:46:14.928963817 +0000 UTC m=+7625.718989417" watchObservedRunningTime="2025-10-06 10:46:14.932133677 +0000 UTC m=+7625.722159257" Oct 06 10:46:15 crc kubenswrapper[4989]: I1006 10:46:15.936974 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:46:15 crc kubenswrapper[4989]: E1006 10:46:15.937920 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:46:30 crc kubenswrapper[4989]: I1006 10:46:30.936827 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:46:30 crc kubenswrapper[4989]: E1006 10:46:30.937699 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.560073 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.563807 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.605430 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.702665 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.702757 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8wk\" (UniqueName: \"kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.703029 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.805816 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.805888 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8wk\" (UniqueName: \"kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.805991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.806335 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.806379 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.834845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8wk\" (UniqueName: \"kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk\") pod \"certified-operators-htzdt\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:31 crc kubenswrapper[4989]: I1006 10:46:31.885873 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:32 crc kubenswrapper[4989]: I1006 10:46:32.464153 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:33 crc kubenswrapper[4989]: I1006 10:46:33.110319 4989 generic.go:334] "Generic (PLEG): container finished" podID="7af87649-0ce8-4855-817f-98d35152429a" containerID="1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f" exitCode=0 Oct 06 10:46:33 crc kubenswrapper[4989]: I1006 10:46:33.110377 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerDied","Data":"1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f"} Oct 06 10:46:33 crc kubenswrapper[4989]: I1006 10:46:33.110408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerStarted","Data":"3bd0aa52ff8ce9951128a46b966ede22495fdf19cf4049456a86ccbe3276e76b"} Oct 06 10:46:34 crc kubenswrapper[4989]: I1006 10:46:34.121408 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerStarted","Data":"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03"} Oct 06 10:46:35 crc kubenswrapper[4989]: I1006 10:46:35.137158 4989 generic.go:334] "Generic (PLEG): container finished" podID="7af87649-0ce8-4855-817f-98d35152429a" containerID="9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03" exitCode=0 Oct 06 10:46:35 crc kubenswrapper[4989]: I1006 10:46:35.137222 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerDied","Data":"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03"} Oct 06 10:46:36 crc kubenswrapper[4989]: I1006 10:46:36.149243 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerStarted","Data":"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d"} Oct 06 10:46:36 crc kubenswrapper[4989]: I1006 10:46:36.177277 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-htzdt" podStartSLOduration=2.74030386 podStartE2EDuration="5.177259383s" podCreationTimestamp="2025-10-06 10:46:31 +0000 UTC" firstStartedPulling="2025-10-06 10:46:33.112273238 +0000 UTC m=+7643.902298818" lastFinishedPulling="2025-10-06 10:46:35.549228761 +0000 UTC m=+7646.339254341" observedRunningTime="2025-10-06 10:46:36.170703625 +0000 UTC m=+7646.960729215" watchObservedRunningTime="2025-10-06 10:46:36.177259383 +0000 UTC m=+7646.967284983" Oct 06 10:46:41 crc kubenswrapper[4989]: I1006 10:46:41.886683 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:41 crc kubenswrapper[4989]: I1006 10:46:41.887199 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:41 crc kubenswrapper[4989]: I1006 10:46:41.955002 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:42 crc kubenswrapper[4989]: I1006 10:46:42.280856 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:42 crc kubenswrapper[4989]: I1006 10:46:42.344787 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.244056 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-htzdt" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="registry-server" containerID="cri-o://e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d" gracePeriod=2 Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.790590 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.910615 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh8wk\" (UniqueName: \"kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk\") pod \"7af87649-0ce8-4855-817f-98d35152429a\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.911141 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content\") pod \"7af87649-0ce8-4855-817f-98d35152429a\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.911176 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities\") pod \"7af87649-0ce8-4855-817f-98d35152429a\" (UID: \"7af87649-0ce8-4855-817f-98d35152429a\") " Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.911976 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities" (OuterVolumeSpecName: "utilities") pod "7af87649-0ce8-4855-817f-98d35152429a" (UID: "7af87649-0ce8-4855-817f-98d35152429a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.916702 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk" (OuterVolumeSpecName: "kube-api-access-jh8wk") pod "7af87649-0ce8-4855-817f-98d35152429a" (UID: "7af87649-0ce8-4855-817f-98d35152429a"). InnerVolumeSpecName "kube-api-access-jh8wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:46:44 crc kubenswrapper[4989]: I1006 10:46:44.969119 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7af87649-0ce8-4855-817f-98d35152429a" (UID: "7af87649-0ce8-4855-817f-98d35152429a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.014189 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh8wk\" (UniqueName: \"kubernetes.io/projected/7af87649-0ce8-4855-817f-98d35152429a-kube-api-access-jh8wk\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.014229 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.014244 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af87649-0ce8-4855-817f-98d35152429a-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.261635 4989 generic.go:334] "Generic (PLEG): container finished" podID="7af87649-0ce8-4855-817f-98d35152429a" containerID="e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d" exitCode=0 Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.261689 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerDied","Data":"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d"} Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.261724 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htzdt" event={"ID":"7af87649-0ce8-4855-817f-98d35152429a","Type":"ContainerDied","Data":"3bd0aa52ff8ce9951128a46b966ede22495fdf19cf4049456a86ccbe3276e76b"} Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.261743 4989 scope.go:117] "RemoveContainer" containerID="e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.261763 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htzdt" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.307585 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.309291 4989 scope.go:117] "RemoveContainer" containerID="9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.316626 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-htzdt"] Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.331596 4989 scope.go:117] "RemoveContainer" containerID="1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.380368 4989 scope.go:117] "RemoveContainer" containerID="e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d" Oct 06 10:46:45 crc kubenswrapper[4989]: E1006 10:46:45.380799 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d\": container with ID starting with e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d not found: ID does not exist" containerID="e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.380842 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d"} err="failed to get container status \"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d\": rpc error: code = NotFound desc = could not find container \"e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d\": container with ID starting with e8a0c20f0734fe939f7ba19c25ac0337c8d3e862972e49c8bcfe9b0b39ba167d not found: ID does not exist" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.380867 4989 scope.go:117] "RemoveContainer" containerID="9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03" Oct 06 10:46:45 crc kubenswrapper[4989]: E1006 10:46:45.381284 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03\": container with ID starting with 9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03 not found: ID does not exist" containerID="9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.381320 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03"} err="failed to get container status \"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03\": rpc error: code = NotFound desc = could not find container \"9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03\": container with ID starting with 9ec7a88946f1ff9e88069e19b8e50be5e1599420f8f2c122bae16636adde0e03 not found: ID does not exist" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.381345 4989 scope.go:117] "RemoveContainer" containerID="1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f" Oct 06 10:46:45 crc kubenswrapper[4989]: E1006 10:46:45.381613 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f\": container with ID starting with 1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f not found: ID does not exist" containerID="1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.381649 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f"} err="failed to get container status \"1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f\": rpc error: code = NotFound desc = could not find container \"1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f\": container with ID starting with 1166b2909ee735877557535434db2bb98f498f878db102e8ba3a9fc596c7337f not found: ID does not exist" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.936455 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:46:45 crc kubenswrapper[4989]: E1006 10:46:45.937136 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:46:45 crc kubenswrapper[4989]: I1006 10:46:45.947560 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af87649-0ce8-4855-817f-98d35152429a" path="/var/lib/kubelet/pods/7af87649-0ce8-4855-817f-98d35152429a/volumes" Oct 06 10:47:00 crc kubenswrapper[4989]: I1006 10:47:00.421709 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" event={"ID":"ee637d80-17d5-4a6c-9734-67f46cdc47f0","Type":"ContainerDied","Data":"3bf2f250819332d82fea007ccbbd2d1cfbde62ecf65edb6c188fe914a824150c"} Oct 06 10:47:00 crc kubenswrapper[4989]: I1006 10:47:00.421644 4989 generic.go:334] "Generic (PLEG): container finished" podID="ee637d80-17d5-4a6c-9734-67f46cdc47f0" containerID="3bf2f250819332d82fea007ccbbd2d1cfbde62ecf65edb6c188fe914a824150c" exitCode=0 Oct 06 10:47:00 crc kubenswrapper[4989]: I1006 10:47:00.940096 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:47:00 crc kubenswrapper[4989]: E1006 10:47:00.940985 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:47:01 crc kubenswrapper[4989]: I1006 10:47:01.941289 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.118213 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph\") pod \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.118512 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key\") pod \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.118541 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory\") pod \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.118576 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl8gn\" (UniqueName: \"kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn\") pod \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\" (UID: \"ee637d80-17d5-4a6c-9734-67f46cdc47f0\") " Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.125575 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph" (OuterVolumeSpecName: "ceph") pod "ee637d80-17d5-4a6c-9734-67f46cdc47f0" (UID: "ee637d80-17d5-4a6c-9734-67f46cdc47f0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.126272 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn" (OuterVolumeSpecName: "kube-api-access-sl8gn") pod "ee637d80-17d5-4a6c-9734-67f46cdc47f0" (UID: "ee637d80-17d5-4a6c-9734-67f46cdc47f0"). InnerVolumeSpecName "kube-api-access-sl8gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.154888 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory" (OuterVolumeSpecName: "inventory") pod "ee637d80-17d5-4a6c-9734-67f46cdc47f0" (UID: "ee637d80-17d5-4a6c-9734-67f46cdc47f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.166673 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee637d80-17d5-4a6c-9734-67f46cdc47f0" (UID: "ee637d80-17d5-4a6c-9734-67f46cdc47f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.221903 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.221934 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.221947 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl8gn\" (UniqueName: \"kubernetes.io/projected/ee637d80-17d5-4a6c-9734-67f46cdc47f0-kube-api-access-sl8gn\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.221960 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ee637d80-17d5-4a6c-9734-67f46cdc47f0-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.447050 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" event={"ID":"ee637d80-17d5-4a6c-9734-67f46cdc47f0","Type":"ContainerDied","Data":"d15c79ad0c680065fb32d66ec99d753df6020d01a2a614a41b212dfd02711313"} Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.447096 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15c79ad0c680065fb32d66ec99d753df6020d01a2a614a41b212dfd02711313" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.447151 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-2xlln" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.568937 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-r29s8"] Oct 06 10:47:02 crc kubenswrapper[4989]: E1006 10:47:02.569362 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="registry-server" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569394 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="registry-server" Oct 06 10:47:02 crc kubenswrapper[4989]: E1006 10:47:02.569415 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="extract-content" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569425 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="extract-content" Oct 06 10:47:02 crc kubenswrapper[4989]: E1006 10:47:02.569448 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee637d80-17d5-4a6c-9734-67f46cdc47f0" containerName="configure-os-openstack-openstack-cell1" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569454 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee637d80-17d5-4a6c-9734-67f46cdc47f0" containerName="configure-os-openstack-openstack-cell1" Oct 06 10:47:02 crc kubenswrapper[4989]: E1006 10:47:02.569462 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="extract-utilities" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569467 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="extract-utilities" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569687 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee637d80-17d5-4a6c-9734-67f46cdc47f0" containerName="configure-os-openstack-openstack-cell1" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.569718 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af87649-0ce8-4855-817f-98d35152429a" containerName="registry-server" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.570569 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.573885 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.573953 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.574896 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.574945 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.586994 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-r29s8"] Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.732807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.733049 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8pr6\" (UniqueName: \"kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.733102 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.733197 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.835445 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.835688 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8pr6\" (UniqueName: \"kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.835780 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.835921 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.843253 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.843379 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.849996 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.859485 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8pr6\" (UniqueName: \"kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6\") pod \"ssh-known-hosts-openstack-r29s8\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:02 crc kubenswrapper[4989]: I1006 10:47:02.895222 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:03 crc kubenswrapper[4989]: I1006 10:47:03.504869 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-r29s8"] Oct 06 10:47:04 crc kubenswrapper[4989]: I1006 10:47:04.471401 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-r29s8" event={"ID":"7eaa911b-7e42-410e-a0f6-36abf0909e7a","Type":"ContainerStarted","Data":"9ed9fd07204c43129b3f5a46d8c91a47810c38a718cd5a89e7c0522e1b0c72a1"} Oct 06 10:47:04 crc kubenswrapper[4989]: I1006 10:47:04.471960 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-r29s8" event={"ID":"7eaa911b-7e42-410e-a0f6-36abf0909e7a","Type":"ContainerStarted","Data":"aa060e4bda082de045e72fb999e48bb32513c945e498675817c204e7b63386fa"} Oct 06 10:47:04 crc kubenswrapper[4989]: I1006 10:47:04.498433 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-r29s8" podStartSLOduration=2.321026889 podStartE2EDuration="2.498416815s" podCreationTimestamp="2025-10-06 10:47:02 +0000 UTC" firstStartedPulling="2025-10-06 10:47:03.501509369 +0000 UTC m=+7674.291534969" lastFinishedPulling="2025-10-06 10:47:03.678899305 +0000 UTC m=+7674.468924895" observedRunningTime="2025-10-06 10:47:04.493041171 +0000 UTC m=+7675.283066751" watchObservedRunningTime="2025-10-06 10:47:04.498416815 +0000 UTC m=+7675.288442395" Oct 06 10:47:12 crc kubenswrapper[4989]: I1006 10:47:12.581977 4989 generic.go:334] "Generic (PLEG): container finished" podID="7eaa911b-7e42-410e-a0f6-36abf0909e7a" containerID="9ed9fd07204c43129b3f5a46d8c91a47810c38a718cd5a89e7c0522e1b0c72a1" exitCode=0 Oct 06 10:47:12 crc kubenswrapper[4989]: I1006 10:47:12.582063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-r29s8" event={"ID":"7eaa911b-7e42-410e-a0f6-36abf0909e7a","Type":"ContainerDied","Data":"9ed9fd07204c43129b3f5a46d8c91a47810c38a718cd5a89e7c0522e1b0c72a1"} Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.040889 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.189903 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1\") pod \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.190022 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8pr6\" (UniqueName: \"kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6\") pod \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.190164 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0\") pod \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.190259 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph\") pod \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\" (UID: \"7eaa911b-7e42-410e-a0f6-36abf0909e7a\") " Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.195981 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph" (OuterVolumeSpecName: "ceph") pod "7eaa911b-7e42-410e-a0f6-36abf0909e7a" (UID: "7eaa911b-7e42-410e-a0f6-36abf0909e7a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.197926 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6" (OuterVolumeSpecName: "kube-api-access-t8pr6") pod "7eaa911b-7e42-410e-a0f6-36abf0909e7a" (UID: "7eaa911b-7e42-410e-a0f6-36abf0909e7a"). InnerVolumeSpecName "kube-api-access-t8pr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.231555 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7eaa911b-7e42-410e-a0f6-36abf0909e7a" (UID: "7eaa911b-7e42-410e-a0f6-36abf0909e7a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.251631 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "7eaa911b-7e42-410e-a0f6-36abf0909e7a" (UID: "7eaa911b-7e42-410e-a0f6-36abf0909e7a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.293324 4989 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.293373 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.293385 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7eaa911b-7e42-410e-a0f6-36abf0909e7a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.293400 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8pr6\" (UniqueName: \"kubernetes.io/projected/7eaa911b-7e42-410e-a0f6-36abf0909e7a-kube-api-access-t8pr6\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.608219 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-r29s8" event={"ID":"7eaa911b-7e42-410e-a0f6-36abf0909e7a","Type":"ContainerDied","Data":"aa060e4bda082de045e72fb999e48bb32513c945e498675817c204e7b63386fa"} Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.608308 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa060e4bda082de045e72fb999e48bb32513c945e498675817c204e7b63386fa" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.608362 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-r29s8" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.686421 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-kb9m6"] Oct 06 10:47:14 crc kubenswrapper[4989]: E1006 10:47:14.686878 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eaa911b-7e42-410e-a0f6-36abf0909e7a" containerName="ssh-known-hosts-openstack" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.686896 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eaa911b-7e42-410e-a0f6-36abf0909e7a" containerName="ssh-known-hosts-openstack" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.687168 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eaa911b-7e42-410e-a0f6-36abf0909e7a" containerName="ssh-known-hosts-openstack" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.688040 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.701221 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.701516 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.701875 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.702724 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.706507 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-kb9m6"] Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.806436 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.806486 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.806773 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56hbg\" (UniqueName: \"kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.807034 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.909242 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56hbg\" (UniqueName: \"kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.909712 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.909814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.909855 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.914558 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.915806 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.918221 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.926471 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56hbg\" (UniqueName: \"kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg\") pod \"run-os-openstack-openstack-cell1-kb9m6\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:14 crc kubenswrapper[4989]: I1006 10:47:14.936287 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:47:14 crc kubenswrapper[4989]: E1006 10:47:14.936706 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:47:15 crc kubenswrapper[4989]: I1006 10:47:15.046320 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:15 crc kubenswrapper[4989]: I1006 10:47:15.725131 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-kb9m6"] Oct 06 10:47:16 crc kubenswrapper[4989]: I1006 10:47:16.640469 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" event={"ID":"889a4162-2caf-4c70-8b84-f406cfb67f4a","Type":"ContainerStarted","Data":"16a6f11c90b3e30569f03785a15a1b320b0ded4488e0cc0c165222045fe1b315"} Oct 06 10:47:16 crc kubenswrapper[4989]: I1006 10:47:16.641030 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" event={"ID":"889a4162-2caf-4c70-8b84-f406cfb67f4a","Type":"ContainerStarted","Data":"47ad6d097daedee372f6f600206dadd4ecb8ec6f5309032029a264c71cbbb7d4"} Oct 06 10:47:16 crc kubenswrapper[4989]: I1006 10:47:16.664200 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" podStartSLOduration=2.464203973 podStartE2EDuration="2.664182545s" podCreationTimestamp="2025-10-06 10:47:14 +0000 UTC" firstStartedPulling="2025-10-06 10:47:15.741909454 +0000 UTC m=+7686.531935034" lastFinishedPulling="2025-10-06 10:47:15.941888016 +0000 UTC m=+7686.731913606" observedRunningTime="2025-10-06 10:47:16.656897066 +0000 UTC m=+7687.446922666" watchObservedRunningTime="2025-10-06 10:47:16.664182545 +0000 UTC m=+7687.454208125" Oct 06 10:47:24 crc kubenswrapper[4989]: I1006 10:47:24.724903 4989 generic.go:334] "Generic (PLEG): container finished" podID="889a4162-2caf-4c70-8b84-f406cfb67f4a" containerID="16a6f11c90b3e30569f03785a15a1b320b0ded4488e0cc0c165222045fe1b315" exitCode=0 Oct 06 10:47:24 crc kubenswrapper[4989]: I1006 10:47:24.724997 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" event={"ID":"889a4162-2caf-4c70-8b84-f406cfb67f4a","Type":"ContainerDied","Data":"16a6f11c90b3e30569f03785a15a1b320b0ded4488e0cc0c165222045fe1b315"} Oct 06 10:47:25 crc kubenswrapper[4989]: I1006 10:47:25.936561 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:47:25 crc kubenswrapper[4989]: E1006 10:47:25.937188 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.228813 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.261361 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph\") pod \"889a4162-2caf-4c70-8b84-f406cfb67f4a\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.261533 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory\") pod \"889a4162-2caf-4c70-8b84-f406cfb67f4a\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.261642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key\") pod \"889a4162-2caf-4c70-8b84-f406cfb67f4a\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.261681 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56hbg\" (UniqueName: \"kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg\") pod \"889a4162-2caf-4c70-8b84-f406cfb67f4a\" (UID: \"889a4162-2caf-4c70-8b84-f406cfb67f4a\") " Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.268959 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph" (OuterVolumeSpecName: "ceph") pod "889a4162-2caf-4c70-8b84-f406cfb67f4a" (UID: "889a4162-2caf-4c70-8b84-f406cfb67f4a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.288989 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg" (OuterVolumeSpecName: "kube-api-access-56hbg") pod "889a4162-2caf-4c70-8b84-f406cfb67f4a" (UID: "889a4162-2caf-4c70-8b84-f406cfb67f4a"). InnerVolumeSpecName "kube-api-access-56hbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.299808 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "889a4162-2caf-4c70-8b84-f406cfb67f4a" (UID: "889a4162-2caf-4c70-8b84-f406cfb67f4a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.299901 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory" (OuterVolumeSpecName: "inventory") pod "889a4162-2caf-4c70-8b84-f406cfb67f4a" (UID: "889a4162-2caf-4c70-8b84-f406cfb67f4a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.364287 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.364510 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.364587 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/889a4162-2caf-4c70-8b84-f406cfb67f4a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.364681 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56hbg\" (UniqueName: \"kubernetes.io/projected/889a4162-2caf-4c70-8b84-f406cfb67f4a-kube-api-access-56hbg\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.741050 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" event={"ID":"889a4162-2caf-4c70-8b84-f406cfb67f4a","Type":"ContainerDied","Data":"47ad6d097daedee372f6f600206dadd4ecb8ec6f5309032029a264c71cbbb7d4"} Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.741085 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47ad6d097daedee372f6f600206dadd4ecb8ec6f5309032029a264c71cbbb7d4" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.741133 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-kb9m6" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.827442 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-8nj2v"] Oct 06 10:47:26 crc kubenswrapper[4989]: E1006 10:47:26.828117 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="889a4162-2caf-4c70-8b84-f406cfb67f4a" containerName="run-os-openstack-openstack-cell1" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.828140 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="889a4162-2caf-4c70-8b84-f406cfb67f4a" containerName="run-os-openstack-openstack-cell1" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.828451 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="889a4162-2caf-4c70-8b84-f406cfb67f4a" containerName="run-os-openstack-openstack-cell1" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.829448 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: E1006 10:47:26.831598 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod889a4162_2caf_4c70_8b84_f406cfb67f4a.slice/crio-47ad6d097daedee372f6f600206dadd4ecb8ec6f5309032029a264c71cbbb7d4\": RecentStats: unable to find data in memory cache]" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.837457 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.839164 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.839614 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.839923 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.849681 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-8nj2v"] Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.877345 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs878\" (UniqueName: \"kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.877460 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.877640 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.877830 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.979369 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs878\" (UniqueName: \"kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.979465 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.979491 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.979551 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.983438 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.983527 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.992346 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:26 crc kubenswrapper[4989]: I1006 10:47:26.996377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs878\" (UniqueName: \"kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878\") pod \"reboot-os-openstack-openstack-cell1-8nj2v\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:27 crc kubenswrapper[4989]: I1006 10:47:27.156614 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:27 crc kubenswrapper[4989]: I1006 10:47:27.811123 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-8nj2v"] Oct 06 10:47:27 crc kubenswrapper[4989]: W1006 10:47:27.818799 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8632b29_9ab1_4376_8b89_6be9d15e7754.slice/crio-5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c WatchSource:0}: Error finding container 5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c: Status 404 returned error can't find the container with id 5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c Oct 06 10:47:28 crc kubenswrapper[4989]: I1006 10:47:28.761515 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" event={"ID":"d8632b29-9ab1-4376-8b89-6be9d15e7754","Type":"ContainerStarted","Data":"f8f02958d16d484d477660b855f83819197ae6eb00257fd79bdca509e29a9463"} Oct 06 10:47:28 crc kubenswrapper[4989]: I1006 10:47:28.762058 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" event={"ID":"d8632b29-9ab1-4376-8b89-6be9d15e7754","Type":"ContainerStarted","Data":"5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c"} Oct 06 10:47:37 crc kubenswrapper[4989]: I1006 10:47:37.936292 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:47:37 crc kubenswrapper[4989]: E1006 10:47:37.936951 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:47:43 crc kubenswrapper[4989]: I1006 10:47:43.915249 4989 generic.go:334] "Generic (PLEG): container finished" podID="d8632b29-9ab1-4376-8b89-6be9d15e7754" containerID="f8f02958d16d484d477660b855f83819197ae6eb00257fd79bdca509e29a9463" exitCode=0 Oct 06 10:47:43 crc kubenswrapper[4989]: I1006 10:47:43.915322 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" event={"ID":"d8632b29-9ab1-4376-8b89-6be9d15e7754","Type":"ContainerDied","Data":"f8f02958d16d484d477660b855f83819197ae6eb00257fd79bdca509e29a9463"} Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.399785 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.513641 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph\") pod \"d8632b29-9ab1-4376-8b89-6be9d15e7754\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.513886 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key\") pod \"d8632b29-9ab1-4376-8b89-6be9d15e7754\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.513997 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs878\" (UniqueName: \"kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878\") pod \"d8632b29-9ab1-4376-8b89-6be9d15e7754\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.514101 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory\") pod \"d8632b29-9ab1-4376-8b89-6be9d15e7754\" (UID: \"d8632b29-9ab1-4376-8b89-6be9d15e7754\") " Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.519766 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878" (OuterVolumeSpecName: "kube-api-access-zs878") pod "d8632b29-9ab1-4376-8b89-6be9d15e7754" (UID: "d8632b29-9ab1-4376-8b89-6be9d15e7754"). InnerVolumeSpecName "kube-api-access-zs878". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.521435 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph" (OuterVolumeSpecName: "ceph") pod "d8632b29-9ab1-4376-8b89-6be9d15e7754" (UID: "d8632b29-9ab1-4376-8b89-6be9d15e7754"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.546873 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory" (OuterVolumeSpecName: "inventory") pod "d8632b29-9ab1-4376-8b89-6be9d15e7754" (UID: "d8632b29-9ab1-4376-8b89-6be9d15e7754"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.552912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8632b29-9ab1-4376-8b89-6be9d15e7754" (UID: "d8632b29-9ab1-4376-8b89-6be9d15e7754"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.618228 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.618271 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.618286 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs878\" (UniqueName: \"kubernetes.io/projected/d8632b29-9ab1-4376-8b89-6be9d15e7754-kube-api-access-zs878\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.618300 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8632b29-9ab1-4376-8b89-6be9d15e7754-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.937466 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.947272 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-8nj2v" event={"ID":"d8632b29-9ab1-4376-8b89-6be9d15e7754","Type":"ContainerDied","Data":"5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c"} Oct 06 10:47:45 crc kubenswrapper[4989]: I1006 10:47:45.947316 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cc72e4c88b741b8da81919d2ecdd6209d46c307685da9dd816f043bea38ac2c" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.019769 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-lfx9n"] Oct 06 10:47:46 crc kubenswrapper[4989]: E1006 10:47:46.020284 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8632b29-9ab1-4376-8b89-6be9d15e7754" containerName="reboot-os-openstack-openstack-cell1" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.020298 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8632b29-9ab1-4376-8b89-6be9d15e7754" containerName="reboot-os-openstack-openstack-cell1" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.020535 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8632b29-9ab1-4376-8b89-6be9d15e7754" containerName="reboot-os-openstack-openstack-cell1" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.021536 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036005 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036038 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036064 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036108 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4kdt\" (UniqueName: \"kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036139 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036174 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036200 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036228 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036264 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036287 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036368 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.036391 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.044286 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-lfx9n"] Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.055119 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.055341 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.055559 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.055742 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138170 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138413 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138491 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138512 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138567 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138585 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4kdt\" (UniqueName: \"kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138712 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138767 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138791 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.138814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.144532 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.144814 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.145127 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.145609 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.145828 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.146664 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.147377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.147742 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.147897 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.147952 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.156214 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.159363 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4kdt\" (UniqueName: \"kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt\") pod \"install-certs-openstack-openstack-cell1-lfx9n\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.391481 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:47:46 crc kubenswrapper[4989]: I1006 10:47:46.959149 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-lfx9n"] Oct 06 10:47:47 crc kubenswrapper[4989]: I1006 10:47:47.975923 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" event={"ID":"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a","Type":"ContainerStarted","Data":"7c7166d5c2cad409044e415d17628f41656499025a9d1f569345bd0a0c9b2d63"} Oct 06 10:47:47 crc kubenswrapper[4989]: I1006 10:47:47.978271 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" event={"ID":"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a","Type":"ContainerStarted","Data":"54bd7d0c518ac1095f6edaa07d04767b49267392d20c2c2d630e2631919ebf33"} Oct 06 10:47:49 crc kubenswrapper[4989]: I1006 10:47:49.947762 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:47:49 crc kubenswrapper[4989]: E1006 10:47:49.948428 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:48:02 crc kubenswrapper[4989]: I1006 10:48:02.935926 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:48:02 crc kubenswrapper[4989]: E1006 10:48:02.936568 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:48:07 crc kubenswrapper[4989]: I1006 10:48:07.181513 4989 generic.go:334] "Generic (PLEG): container finished" podID="d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" containerID="7c7166d5c2cad409044e415d17628f41656499025a9d1f569345bd0a0c9b2d63" exitCode=0 Oct 06 10:48:07 crc kubenswrapper[4989]: I1006 10:48:07.181622 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" event={"ID":"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a","Type":"ContainerDied","Data":"7c7166d5c2cad409044e415d17628f41656499025a9d1f569345bd0a0c9b2d63"} Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.622234 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758613 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758687 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758774 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758851 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758877 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.758980 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759007 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759041 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759067 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759128 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759186 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4kdt\" (UniqueName: \"kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.759308 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle\") pod \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\" (UID: \"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a\") " Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.765399 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.765420 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.765446 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.767169 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt" (OuterVolumeSpecName: "kube-api-access-x4kdt") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "kube-api-access-x4kdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.767243 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.767354 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.767462 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.767473 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.768586 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph" (OuterVolumeSpecName: "ceph") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.768973 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.791336 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory" (OuterVolumeSpecName: "inventory") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.814563 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" (UID: "d3cacb3e-da1d-4bb8-b60a-890f8a8f935a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863494 4989 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863544 4989 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863565 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863590 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863608 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863625 4989 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863642 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863684 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863700 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863716 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863734 4989 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:08 crc kubenswrapper[4989]: I1006 10:48:08.863752 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4kdt\" (UniqueName: \"kubernetes.io/projected/d3cacb3e-da1d-4bb8-b60a-890f8a8f935a-kube-api-access-x4kdt\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.203053 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" event={"ID":"d3cacb3e-da1d-4bb8-b60a-890f8a8f935a","Type":"ContainerDied","Data":"54bd7d0c518ac1095f6edaa07d04767b49267392d20c2c2d630e2631919ebf33"} Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.203101 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54bd7d0c518ac1095f6edaa07d04767b49267392d20c2c2d630e2631919ebf33" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.203159 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-lfx9n" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.287129 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-qmlnc"] Oct 06 10:48:09 crc kubenswrapper[4989]: E1006 10:48:09.288082 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" containerName="install-certs-openstack-openstack-cell1" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.288106 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" containerName="install-certs-openstack-openstack-cell1" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.288395 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3cacb3e-da1d-4bb8-b60a-890f8a8f935a" containerName="install-certs-openstack-openstack-cell1" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.289492 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.291933 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.292225 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.299062 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.299365 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.302147 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-qmlnc"] Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.375004 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.375061 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.375099 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-strd4\" (UniqueName: \"kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.375424 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.477365 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.477511 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.477557 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.477602 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-strd4\" (UniqueName: \"kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.480867 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.480917 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.485522 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.495716 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-strd4\" (UniqueName: \"kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4\") pod \"ceph-client-openstack-openstack-cell1-qmlnc\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:09 crc kubenswrapper[4989]: I1006 10:48:09.623256 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:10 crc kubenswrapper[4989]: I1006 10:48:10.245467 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-qmlnc"] Oct 06 10:48:10 crc kubenswrapper[4989]: I1006 10:48:10.419273 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:48:11 crc kubenswrapper[4989]: I1006 10:48:11.228055 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" event={"ID":"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d","Type":"ContainerStarted","Data":"27b22037c9db87d962c02a6597479f8f7b96b88bebdc009ed54119b4e114c911"} Oct 06 10:48:11 crc kubenswrapper[4989]: I1006 10:48:11.228412 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" event={"ID":"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d","Type":"ContainerStarted","Data":"0e042ffe1c63a792590f6aee3f9e050eec62b3b078b7561053da7bd2800d3b70"} Oct 06 10:48:11 crc kubenswrapper[4989]: I1006 10:48:11.248453 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" podStartSLOduration=2.090913722 podStartE2EDuration="2.248414769s" podCreationTimestamp="2025-10-06 10:48:09 +0000 UTC" firstStartedPulling="2025-10-06 10:48:10.257332689 +0000 UTC m=+7741.047358269" lastFinishedPulling="2025-10-06 10:48:10.414833736 +0000 UTC m=+7741.204859316" observedRunningTime="2025-10-06 10:48:11.245292119 +0000 UTC m=+7742.035317719" watchObservedRunningTime="2025-10-06 10:48:11.248414769 +0000 UTC m=+7742.038440359" Oct 06 10:48:14 crc kubenswrapper[4989]: I1006 10:48:14.937800 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:48:14 crc kubenswrapper[4989]: E1006 10:48:14.939287 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:48:16 crc kubenswrapper[4989]: I1006 10:48:16.272756 4989 generic.go:334] "Generic (PLEG): container finished" podID="5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" containerID="27b22037c9db87d962c02a6597479f8f7b96b88bebdc009ed54119b4e114c911" exitCode=0 Oct 06 10:48:16 crc kubenswrapper[4989]: I1006 10:48:16.272897 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" event={"ID":"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d","Type":"ContainerDied","Data":"27b22037c9db87d962c02a6597479f8f7b96b88bebdc009ed54119b4e114c911"} Oct 06 10:48:17 crc kubenswrapper[4989]: I1006 10:48:17.861430 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.012893 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph\") pod \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.013060 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key\") pod \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.013120 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-strd4\" (UniqueName: \"kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4\") pod \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.013250 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory\") pod \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\" (UID: \"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d\") " Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.022388 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph" (OuterVolumeSpecName: "ceph") pod "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" (UID: "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.024854 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4" (OuterVolumeSpecName: "kube-api-access-strd4") pod "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" (UID: "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d"). InnerVolumeSpecName "kube-api-access-strd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.065269 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory" (OuterVolumeSpecName: "inventory") pod "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" (UID: "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.070557 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" (UID: "5fba31cf-6e75-4ffa-85e5-7d789ad6f92d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.116798 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.116859 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.116875 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-strd4\" (UniqueName: \"kubernetes.io/projected/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-kube-api-access-strd4\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.116887 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fba31cf-6e75-4ffa-85e5-7d789ad6f92d-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.297235 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" event={"ID":"5fba31cf-6e75-4ffa-85e5-7d789ad6f92d","Type":"ContainerDied","Data":"0e042ffe1c63a792590f6aee3f9e050eec62b3b078b7561053da7bd2800d3b70"} Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.297483 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e042ffe1c63a792590f6aee3f9e050eec62b3b078b7561053da7bd2800d3b70" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.297269 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-qmlnc" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.378893 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-fw472"] Oct 06 10:48:18 crc kubenswrapper[4989]: E1006 10:48:18.394172 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" containerName="ceph-client-openstack-openstack-cell1" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.394214 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" containerName="ceph-client-openstack-openstack-cell1" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.394528 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fba31cf-6e75-4ffa-85e5-7d789ad6f92d" containerName="ceph-client-openstack-openstack-cell1" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.395494 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.402807 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.407846 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.408219 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.408916 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.409210 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.440960 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-fw472"] Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.523791 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.523874 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.523910 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.524019 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.524065 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.524207 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbh88\" (UniqueName: \"kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.627605 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.627713 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.627758 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.627789 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.627852 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbh88\" (UniqueName: \"kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.628029 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.632848 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.634416 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.638379 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.639055 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.641754 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.667914 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbh88\" (UniqueName: \"kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88\") pod \"ovn-openstack-openstack-cell1-fw472\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:18 crc kubenswrapper[4989]: I1006 10:48:18.718565 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:48:19 crc kubenswrapper[4989]: I1006 10:48:19.301730 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-fw472"] Oct 06 10:48:20 crc kubenswrapper[4989]: I1006 10:48:20.318552 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-fw472" event={"ID":"e83646ac-1df3-40bc-8706-53b5e1de675f","Type":"ContainerStarted","Data":"4de73b935b67b8ae3575db6b60a903bd01d30a9a98f8d4924a6a6266ef2c9a07"} Oct 06 10:48:20 crc kubenswrapper[4989]: I1006 10:48:20.318900 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-fw472" event={"ID":"e83646ac-1df3-40bc-8706-53b5e1de675f","Type":"ContainerStarted","Data":"9f56e49167b928d67859cfc77a2a2e3fe9af760363d903492b48bf0bbaee7c09"} Oct 06 10:48:20 crc kubenswrapper[4989]: I1006 10:48:20.343582 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-fw472" podStartSLOduration=2.214311345 podStartE2EDuration="2.343564444s" podCreationTimestamp="2025-10-06 10:48:18 +0000 UTC" firstStartedPulling="2025-10-06 10:48:19.312175751 +0000 UTC m=+7750.102201341" lastFinishedPulling="2025-10-06 10:48:19.44142886 +0000 UTC m=+7750.231454440" observedRunningTime="2025-10-06 10:48:20.33295891 +0000 UTC m=+7751.122984490" watchObservedRunningTime="2025-10-06 10:48:20.343564444 +0000 UTC m=+7751.133590024" Oct 06 10:48:26 crc kubenswrapper[4989]: I1006 10:48:26.936680 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:48:26 crc kubenswrapper[4989]: E1006 10:48:26.937585 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:48:40 crc kubenswrapper[4989]: I1006 10:48:40.937178 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:48:41 crc kubenswrapper[4989]: I1006 10:48:41.564313 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42"} Oct 06 10:49:27 crc kubenswrapper[4989]: I1006 10:49:27.104539 4989 generic.go:334] "Generic (PLEG): container finished" podID="e83646ac-1df3-40bc-8706-53b5e1de675f" containerID="4de73b935b67b8ae3575db6b60a903bd01d30a9a98f8d4924a6a6266ef2c9a07" exitCode=0 Oct 06 10:49:27 crc kubenswrapper[4989]: I1006 10:49:27.104695 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-fw472" event={"ID":"e83646ac-1df3-40bc-8706-53b5e1de675f","Type":"ContainerDied","Data":"4de73b935b67b8ae3575db6b60a903bd01d30a9a98f8d4924a6a6266ef2c9a07"} Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.686162 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776380 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776479 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776545 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbh88\" (UniqueName: \"kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776614 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.776709 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key\") pod \"e83646ac-1df3-40bc-8706-53b5e1de675f\" (UID: \"e83646ac-1df3-40bc-8706-53b5e1de675f\") " Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.786137 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph" (OuterVolumeSpecName: "ceph") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.786541 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88" (OuterVolumeSpecName: "kube-api-access-hbh88") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "kube-api-access-hbh88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.786888 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.817418 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory" (OuterVolumeSpecName: "inventory") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.818977 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.826224 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e83646ac-1df3-40bc-8706-53b5e1de675f" (UID: "e83646ac-1df3-40bc-8706-53b5e1de675f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878631 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878681 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878695 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878709 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbh88\" (UniqueName: \"kubernetes.io/projected/e83646ac-1df3-40bc-8706-53b5e1de675f-kube-api-access-hbh88\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878721 4989 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e83646ac-1df3-40bc-8706-53b5e1de675f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:28 crc kubenswrapper[4989]: I1006 10:49:28.878731 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e83646ac-1df3-40bc-8706-53b5e1de675f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.156812 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-fw472" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.156833 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-fw472" event={"ID":"e83646ac-1df3-40bc-8706-53b5e1de675f","Type":"ContainerDied","Data":"9f56e49167b928d67859cfc77a2a2e3fe9af760363d903492b48bf0bbaee7c09"} Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.157015 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f56e49167b928d67859cfc77a2a2e3fe9af760363d903492b48bf0bbaee7c09" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.257230 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-89hhj"] Oct 06 10:49:29 crc kubenswrapper[4989]: E1006 10:49:29.257883 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83646ac-1df3-40bc-8706-53b5e1de675f" containerName="ovn-openstack-openstack-cell1" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.257905 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83646ac-1df3-40bc-8706-53b5e1de675f" containerName="ovn-openstack-openstack-cell1" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.258203 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83646ac-1df3-40bc-8706-53b5e1de675f" containerName="ovn-openstack-openstack-cell1" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.259259 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.264127 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.264229 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.264127 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.264591 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.264787 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.266348 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.296696 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-89hhj"] Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297322 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297440 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297595 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297682 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4bn8\" (UniqueName: \"kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297721 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297779 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.297832 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400103 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400170 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4bn8\" (UniqueName: \"kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400202 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400246 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400285 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400381 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.400448 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.405333 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.405669 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.406426 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.407020 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.409226 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.419250 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.424442 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4bn8\" (UniqueName: \"kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8\") pod \"neutron-metadata-openstack-openstack-cell1-89hhj\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:29 crc kubenswrapper[4989]: I1006 10:49:29.599025 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:49:30 crc kubenswrapper[4989]: I1006 10:49:30.260585 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-89hhj"] Oct 06 10:49:31 crc kubenswrapper[4989]: I1006 10:49:31.186438 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" event={"ID":"1906849c-8f07-411b-b7dc-974e982b96f8","Type":"ContainerStarted","Data":"bea527178a468b7fc59e1df6cde4b4199b6c9ddd8acb9b098cb907eb16430578"} Oct 06 10:49:31 crc kubenswrapper[4989]: I1006 10:49:31.186891 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" event={"ID":"1906849c-8f07-411b-b7dc-974e982b96f8","Type":"ContainerStarted","Data":"8e9c1d1feeed10610fd2e108307fdc213d00fdc5989d9428306096090f0e114d"} Oct 06 10:49:31 crc kubenswrapper[4989]: I1006 10:49:31.212790 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" podStartSLOduration=2.025556661 podStartE2EDuration="2.212775558s" podCreationTimestamp="2025-10-06 10:49:29 +0000 UTC" firstStartedPulling="2025-10-06 10:49:30.277046242 +0000 UTC m=+7821.067071832" lastFinishedPulling="2025-10-06 10:49:30.464265149 +0000 UTC m=+7821.254290729" observedRunningTime="2025-10-06 10:49:31.210114401 +0000 UTC m=+7822.000139981" watchObservedRunningTime="2025-10-06 10:49:31.212775558 +0000 UTC m=+7822.002801138" Oct 06 10:50:26 crc kubenswrapper[4989]: I1006 10:50:26.778167 4989 generic.go:334] "Generic (PLEG): container finished" podID="1906849c-8f07-411b-b7dc-974e982b96f8" containerID="bea527178a468b7fc59e1df6cde4b4199b6c9ddd8acb9b098cb907eb16430578" exitCode=0 Oct 06 10:50:26 crc kubenswrapper[4989]: I1006 10:50:26.778287 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" event={"ID":"1906849c-8f07-411b-b7dc-974e982b96f8","Type":"ContainerDied","Data":"bea527178a468b7fc59e1df6cde4b4199b6c9ddd8acb9b098cb907eb16430578"} Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.206699 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.327754 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328109 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328135 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328211 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328251 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4bn8\" (UniqueName: \"kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328331 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.328371 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory\") pod \"1906849c-8f07-411b-b7dc-974e982b96f8\" (UID: \"1906849c-8f07-411b-b7dc-974e982b96f8\") " Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.333893 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.334939 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8" (OuterVolumeSpecName: "kube-api-access-x4bn8") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "kube-api-access-x4bn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.335828 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph" (OuterVolumeSpecName: "ceph") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.360780 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.361165 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory" (OuterVolumeSpecName: "inventory") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.362328 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.362758 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "1906849c-8f07-411b-b7dc-974e982b96f8" (UID: "1906849c-8f07-411b-b7dc-974e982b96f8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430793 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430828 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430839 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430849 4989 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430861 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430871 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1906849c-8f07-411b-b7dc-974e982b96f8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.430880 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4bn8\" (UniqueName: \"kubernetes.io/projected/1906849c-8f07-411b-b7dc-974e982b96f8-kube-api-access-x4bn8\") on node \"crc\" DevicePath \"\"" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.798082 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" event={"ID":"1906849c-8f07-411b-b7dc-974e982b96f8","Type":"ContainerDied","Data":"8e9c1d1feeed10610fd2e108307fdc213d00fdc5989d9428306096090f0e114d"} Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.798127 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e9c1d1feeed10610fd2e108307fdc213d00fdc5989d9428306096090f0e114d" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.798155 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-89hhj" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.881339 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-kr879"] Oct 06 10:50:28 crc kubenswrapper[4989]: E1006 10:50:28.881768 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1906849c-8f07-411b-b7dc-974e982b96f8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.881786 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1906849c-8f07-411b-b7dc-974e982b96f8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.882065 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1906849c-8f07-411b-b7dc-974e982b96f8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.882819 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.886134 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.886322 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.886389 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.886460 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.887801 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.898002 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-kr879"] Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942374 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nvds\" (UniqueName: \"kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942483 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942514 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942734 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942782 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:28 crc kubenswrapper[4989]: I1006 10:50:28.942813 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.044644 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.044715 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.044894 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.044951 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.044978 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.045029 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nvds\" (UniqueName: \"kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.048062 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.048067 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.049491 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.051993 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.052143 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.062397 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nvds\" (UniqueName: \"kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds\") pod \"libvirt-openstack-openstack-cell1-kr879\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.201411 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.763841 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-kr879"] Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.772603 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:50:29 crc kubenswrapper[4989]: I1006 10:50:29.809796 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-kr879" event={"ID":"348d43fe-9095-4a19-b9ac-b3dc9e5da427","Type":"ContainerStarted","Data":"dae2743c3d8ba9334dfd5f9f49733a33b0ab83a6cfc848195919ab06244b659e"} Oct 06 10:50:30 crc kubenswrapper[4989]: I1006 10:50:30.822220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-kr879" event={"ID":"348d43fe-9095-4a19-b9ac-b3dc9e5da427","Type":"ContainerStarted","Data":"68a30ab6c71a50e559aa1a7cf69f5965544f0c663110600a0367f74bcd066721"} Oct 06 10:50:30 crc kubenswrapper[4989]: I1006 10:50:30.846786 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-kr879" podStartSLOduration=2.691822864 podStartE2EDuration="2.846763748s" podCreationTimestamp="2025-10-06 10:50:28 +0000 UTC" firstStartedPulling="2025-10-06 10:50:29.772211639 +0000 UTC m=+7880.562237229" lastFinishedPulling="2025-10-06 10:50:29.927152523 +0000 UTC m=+7880.717178113" observedRunningTime="2025-10-06 10:50:30.836490484 +0000 UTC m=+7881.626516074" watchObservedRunningTime="2025-10-06 10:50:30.846763748 +0000 UTC m=+7881.636789338" Oct 06 10:51:03 crc kubenswrapper[4989]: I1006 10:51:03.936617 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:51:03 crc kubenswrapper[4989]: I1006 10:51:03.937340 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.634147 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.636790 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.646990 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.731182 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.731255 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j429g\" (UniqueName: \"kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.731291 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.833817 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.833925 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j429g\" (UniqueName: \"kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.833969 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.834428 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.834688 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.859728 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j429g\" (UniqueName: \"kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g\") pod \"redhat-marketplace-k2gcp\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:16 crc kubenswrapper[4989]: I1006 10:51:16.971191 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:17 crc kubenswrapper[4989]: I1006 10:51:17.455577 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:18 crc kubenswrapper[4989]: I1006 10:51:18.356934 4989 generic.go:334] "Generic (PLEG): container finished" podID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerID="dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2" exitCode=0 Oct 06 10:51:18 crc kubenswrapper[4989]: I1006 10:51:18.357085 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerDied","Data":"dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2"} Oct 06 10:51:18 crc kubenswrapper[4989]: I1006 10:51:18.357279 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerStarted","Data":"73ab956f07d02d6d66a089ce6f4539c9c893967a0b43275003e28e68c553ecdd"} Oct 06 10:51:19 crc kubenswrapper[4989]: I1006 10:51:19.367332 4989 generic.go:334] "Generic (PLEG): container finished" podID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerID="b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d" exitCode=0 Oct 06 10:51:19 crc kubenswrapper[4989]: I1006 10:51:19.367389 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerDied","Data":"b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d"} Oct 06 10:51:20 crc kubenswrapper[4989]: I1006 10:51:20.379095 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerStarted","Data":"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49"} Oct 06 10:51:20 crc kubenswrapper[4989]: I1006 10:51:20.400085 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2gcp" podStartSLOduration=2.788249482 podStartE2EDuration="4.400063873s" podCreationTimestamp="2025-10-06 10:51:16 +0000 UTC" firstStartedPulling="2025-10-06 10:51:18.359352169 +0000 UTC m=+7929.149377759" lastFinishedPulling="2025-10-06 10:51:19.97116657 +0000 UTC m=+7930.761192150" observedRunningTime="2025-10-06 10:51:20.393171266 +0000 UTC m=+7931.183196846" watchObservedRunningTime="2025-10-06 10:51:20.400063873 +0000 UTC m=+7931.190089453" Oct 06 10:51:26 crc kubenswrapper[4989]: I1006 10:51:26.971741 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:26 crc kubenswrapper[4989]: I1006 10:51:26.972210 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:27 crc kubenswrapper[4989]: I1006 10:51:27.020302 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:27 crc kubenswrapper[4989]: I1006 10:51:27.524289 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:27 crc kubenswrapper[4989]: I1006 10:51:27.571553 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:29 crc kubenswrapper[4989]: I1006 10:51:29.493056 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k2gcp" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="registry-server" containerID="cri-o://b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49" gracePeriod=2 Oct 06 10:51:29 crc kubenswrapper[4989]: I1006 10:51:29.965700 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.126137 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities\") pod \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.126271 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content\") pod \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.126318 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j429g\" (UniqueName: \"kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g\") pod \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\" (UID: \"0b611bed-5bd4-4d0f-9d90-2924172bdf3f\") " Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.127242 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities" (OuterVolumeSpecName: "utilities") pod "0b611bed-5bd4-4d0f-9d90-2924172bdf3f" (UID: "0b611bed-5bd4-4d0f-9d90-2924172bdf3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.127350 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.132456 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g" (OuterVolumeSpecName: "kube-api-access-j429g") pod "0b611bed-5bd4-4d0f-9d90-2924172bdf3f" (UID: "0b611bed-5bd4-4d0f-9d90-2924172bdf3f"). InnerVolumeSpecName "kube-api-access-j429g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.141819 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b611bed-5bd4-4d0f-9d90-2924172bdf3f" (UID: "0b611bed-5bd4-4d0f-9d90-2924172bdf3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.229289 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.229317 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j429g\" (UniqueName: \"kubernetes.io/projected/0b611bed-5bd4-4d0f-9d90-2924172bdf3f-kube-api-access-j429g\") on node \"crc\" DevicePath \"\"" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.504195 4989 generic.go:334] "Generic (PLEG): container finished" podID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerID="b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49" exitCode=0 Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.504257 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerDied","Data":"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49"} Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.504511 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2gcp" event={"ID":"0b611bed-5bd4-4d0f-9d90-2924172bdf3f","Type":"ContainerDied","Data":"73ab956f07d02d6d66a089ce6f4539c9c893967a0b43275003e28e68c553ecdd"} Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.504537 4989 scope.go:117] "RemoveContainer" containerID="b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.504299 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2gcp" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.528624 4989 scope.go:117] "RemoveContainer" containerID="b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.543614 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.554389 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2gcp"] Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.566486 4989 scope.go:117] "RemoveContainer" containerID="dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.618840 4989 scope.go:117] "RemoveContainer" containerID="b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49" Oct 06 10:51:30 crc kubenswrapper[4989]: E1006 10:51:30.620397 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49\": container with ID starting with b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49 not found: ID does not exist" containerID="b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.620449 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49"} err="failed to get container status \"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49\": rpc error: code = NotFound desc = could not find container \"b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49\": container with ID starting with b4f850bcb1a573afab09b2545b9ac56ae08347aad22135ca24d5268119eb1b49 not found: ID does not exist" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.620479 4989 scope.go:117] "RemoveContainer" containerID="b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d" Oct 06 10:51:30 crc kubenswrapper[4989]: E1006 10:51:30.620872 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d\": container with ID starting with b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d not found: ID does not exist" containerID="b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.620914 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d"} err="failed to get container status \"b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d\": rpc error: code = NotFound desc = could not find container \"b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d\": container with ID starting with b5fd91ddda965e428a250d1d0ca871ae552d89ebd9e79644c9078be474638d5d not found: ID does not exist" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.620941 4989 scope.go:117] "RemoveContainer" containerID="dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2" Oct 06 10:51:30 crc kubenswrapper[4989]: E1006 10:51:30.621259 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2\": container with ID starting with dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2 not found: ID does not exist" containerID="dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2" Oct 06 10:51:30 crc kubenswrapper[4989]: I1006 10:51:30.621298 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2"} err="failed to get container status \"dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2\": rpc error: code = NotFound desc = could not find container \"dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2\": container with ID starting with dd84599bb40a8169007050da1a565f62540db327a72e6dfed8e48b77785621a2 not found: ID does not exist" Oct 06 10:51:31 crc kubenswrapper[4989]: I1006 10:51:31.952949 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" path="/var/lib/kubelet/pods/0b611bed-5bd4-4d0f-9d90-2924172bdf3f/volumes" Oct 06 10:51:33 crc kubenswrapper[4989]: I1006 10:51:33.935312 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:51:33 crc kubenswrapper[4989]: I1006 10:51:33.935678 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:52:03 crc kubenswrapper[4989]: I1006 10:52:03.937686 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:52:03 crc kubenswrapper[4989]: I1006 10:52:03.938330 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:52:03 crc kubenswrapper[4989]: I1006 10:52:03.947830 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:52:03 crc kubenswrapper[4989]: I1006 10:52:03.948325 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:52:03 crc kubenswrapper[4989]: I1006 10:52:03.948389 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42" gracePeriod=600 Oct 06 10:52:04 crc kubenswrapper[4989]: I1006 10:52:04.896103 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42" exitCode=0 Oct 06 10:52:04 crc kubenswrapper[4989]: I1006 10:52:04.896187 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42"} Oct 06 10:52:04 crc kubenswrapper[4989]: I1006 10:52:04.896629 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb"} Oct 06 10:52:04 crc kubenswrapper[4989]: I1006 10:52:04.896667 4989 scope.go:117] "RemoveContainer" containerID="d9b192fa9b5a3d74c36e43dfa4e37c9aaa97a05c4fe800a777e8c1abcce6d13e" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.440015 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:52:58 crc kubenswrapper[4989]: E1006 10:52:58.441317 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="extract-utilities" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.441336 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="extract-utilities" Oct 06 10:52:58 crc kubenswrapper[4989]: E1006 10:52:58.441364 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="extract-content" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.441372 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="extract-content" Oct 06 10:52:58 crc kubenswrapper[4989]: E1006 10:52:58.441410 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="registry-server" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.441418 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="registry-server" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.441798 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b611bed-5bd4-4d0f-9d90-2924172bdf3f" containerName="registry-server" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.446088 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.468355 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.561665 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.561753 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6krsw\" (UniqueName: \"kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.561788 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.664445 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.664505 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6krsw\" (UniqueName: \"kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.664527 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.665078 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.665084 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.688625 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6krsw\" (UniqueName: \"kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw\") pod \"community-operators-7lnt5\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:58 crc kubenswrapper[4989]: I1006 10:52:58.785900 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:52:59 crc kubenswrapper[4989]: I1006 10:52:59.309593 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:52:59 crc kubenswrapper[4989]: W1006 10:52:59.316665 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaba16934_2d2e_4227_bd91_6c156d8347d1.slice/crio-15f3590912fdbf086e5f554902966cb639288e9905cb0f5a288eb964269eee07 WatchSource:0}: Error finding container 15f3590912fdbf086e5f554902966cb639288e9905cb0f5a288eb964269eee07: Status 404 returned error can't find the container with id 15f3590912fdbf086e5f554902966cb639288e9905cb0f5a288eb964269eee07 Oct 06 10:52:59 crc kubenswrapper[4989]: I1006 10:52:59.563420 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerID="d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe" exitCode=0 Oct 06 10:52:59 crc kubenswrapper[4989]: I1006 10:52:59.563461 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerDied","Data":"d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe"} Oct 06 10:52:59 crc kubenswrapper[4989]: I1006 10:52:59.563485 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerStarted","Data":"15f3590912fdbf086e5f554902966cb639288e9905cb0f5a288eb964269eee07"} Oct 06 10:53:00 crc kubenswrapper[4989]: I1006 10:53:00.574624 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerStarted","Data":"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed"} Oct 06 10:53:01 crc kubenswrapper[4989]: I1006 10:53:01.591561 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerID="577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed" exitCode=0 Oct 06 10:53:01 crc kubenswrapper[4989]: I1006 10:53:01.591640 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerDied","Data":"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed"} Oct 06 10:53:03 crc kubenswrapper[4989]: I1006 10:53:03.617771 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerStarted","Data":"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84"} Oct 06 10:53:03 crc kubenswrapper[4989]: I1006 10:53:03.641563 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7lnt5" podStartSLOduration=2.855869325 podStartE2EDuration="5.641541536s" podCreationTimestamp="2025-10-06 10:52:58 +0000 UTC" firstStartedPulling="2025-10-06 10:52:59.565352517 +0000 UTC m=+8030.355378097" lastFinishedPulling="2025-10-06 10:53:02.351024688 +0000 UTC m=+8033.141050308" observedRunningTime="2025-10-06 10:53:03.636891762 +0000 UTC m=+8034.426917352" watchObservedRunningTime="2025-10-06 10:53:03.641541536 +0000 UTC m=+8034.431567116" Oct 06 10:53:08 crc kubenswrapper[4989]: I1006 10:53:08.786762 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:08 crc kubenswrapper[4989]: I1006 10:53:08.787509 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:08 crc kubenswrapper[4989]: I1006 10:53:08.892627 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:09 crc kubenswrapper[4989]: I1006 10:53:09.770941 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:09 crc kubenswrapper[4989]: I1006 10:53:09.848317 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:53:11 crc kubenswrapper[4989]: I1006 10:53:11.709162 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7lnt5" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="registry-server" containerID="cri-o://e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84" gracePeriod=2 Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.292554 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.409046 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities\") pod \"aba16934-2d2e-4227-bd91-6c156d8347d1\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.409266 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content\") pod \"aba16934-2d2e-4227-bd91-6c156d8347d1\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.409366 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6krsw\" (UniqueName: \"kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw\") pod \"aba16934-2d2e-4227-bd91-6c156d8347d1\" (UID: \"aba16934-2d2e-4227-bd91-6c156d8347d1\") " Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.411594 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities" (OuterVolumeSpecName: "utilities") pod "aba16934-2d2e-4227-bd91-6c156d8347d1" (UID: "aba16934-2d2e-4227-bd91-6c156d8347d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.416997 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw" (OuterVolumeSpecName: "kube-api-access-6krsw") pod "aba16934-2d2e-4227-bd91-6c156d8347d1" (UID: "aba16934-2d2e-4227-bd91-6c156d8347d1"). InnerVolumeSpecName "kube-api-access-6krsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.482945 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aba16934-2d2e-4227-bd91-6c156d8347d1" (UID: "aba16934-2d2e-4227-bd91-6c156d8347d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.512841 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.512886 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba16934-2d2e-4227-bd91-6c156d8347d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.512901 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6krsw\" (UniqueName: \"kubernetes.io/projected/aba16934-2d2e-4227-bd91-6c156d8347d1-kube-api-access-6krsw\") on node \"crc\" DevicePath \"\"" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.723924 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerID="e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84" exitCode=0 Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.724021 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerDied","Data":"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84"} Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.724048 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7lnt5" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.724063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7lnt5" event={"ID":"aba16934-2d2e-4227-bd91-6c156d8347d1","Type":"ContainerDied","Data":"15f3590912fdbf086e5f554902966cb639288e9905cb0f5a288eb964269eee07"} Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.724089 4989 scope.go:117] "RemoveContainer" containerID="e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.768367 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.778061 4989 scope.go:117] "RemoveContainer" containerID="577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.778751 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7lnt5"] Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.807094 4989 scope.go:117] "RemoveContainer" containerID="d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.892607 4989 scope.go:117] "RemoveContainer" containerID="e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84" Oct 06 10:53:12 crc kubenswrapper[4989]: E1006 10:53:12.893346 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84\": container with ID starting with e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84 not found: ID does not exist" containerID="e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.893395 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84"} err="failed to get container status \"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84\": rpc error: code = NotFound desc = could not find container \"e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84\": container with ID starting with e8a67c0ba894c6f3967d1c6b48e0468a3b3b69081305cbbab7f2c0ddbe4d8d84 not found: ID does not exist" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.893424 4989 scope.go:117] "RemoveContainer" containerID="577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed" Oct 06 10:53:12 crc kubenswrapper[4989]: E1006 10:53:12.893826 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed\": container with ID starting with 577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed not found: ID does not exist" containerID="577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.893848 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed"} err="failed to get container status \"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed\": rpc error: code = NotFound desc = could not find container \"577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed\": container with ID starting with 577d846b2c09c7086e33df13b905a2f94dde31364eba146faa33c66b0ebf04ed not found: ID does not exist" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.893862 4989 scope.go:117] "RemoveContainer" containerID="d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe" Oct 06 10:53:12 crc kubenswrapper[4989]: E1006 10:53:12.894129 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe\": container with ID starting with d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe not found: ID does not exist" containerID="d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe" Oct 06 10:53:12 crc kubenswrapper[4989]: I1006 10:53:12.894151 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe"} err="failed to get container status \"d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe\": rpc error: code = NotFound desc = could not find container \"d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe\": container with ID starting with d4e936685da13877c845f1c75f89f91f2e13f64546780bb2c2dfc4adfeb1aabe not found: ID does not exist" Oct 06 10:53:13 crc kubenswrapper[4989]: I1006 10:53:13.951881 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" path="/var/lib/kubelet/pods/aba16934-2d2e-4227-bd91-6c156d8347d1/volumes" Oct 06 10:54:33 crc kubenswrapper[4989]: I1006 10:54:33.935543 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:54:33 crc kubenswrapper[4989]: I1006 10:54:33.936112 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:55:03 crc kubenswrapper[4989]: I1006 10:55:03.935509 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:55:03 crc kubenswrapper[4989]: I1006 10:55:03.936215 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:55:10 crc kubenswrapper[4989]: E1006 10:55:10.761274 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod348d43fe_9095_4a19_b9ac_b3dc9e5da427.slice/crio-conmon-68a30ab6c71a50e559aa1a7cf69f5965544f0c663110600a0367f74bcd066721.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod348d43fe_9095_4a19_b9ac_b3dc9e5da427.slice/crio-68a30ab6c71a50e559aa1a7cf69f5965544f0c663110600a0367f74bcd066721.scope\": RecentStats: unable to find data in memory cache]" Oct 06 10:55:11 crc kubenswrapper[4989]: I1006 10:55:11.117233 4989 generic.go:334] "Generic (PLEG): container finished" podID="348d43fe-9095-4a19-b9ac-b3dc9e5da427" containerID="68a30ab6c71a50e559aa1a7cf69f5965544f0c663110600a0367f74bcd066721" exitCode=0 Oct 06 10:55:11 crc kubenswrapper[4989]: I1006 10:55:11.117291 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-kr879" event={"ID":"348d43fe-9095-4a19-b9ac-b3dc9e5da427","Type":"ContainerDied","Data":"68a30ab6c71a50e559aa1a7cf69f5965544f0c663110600a0367f74bcd066721"} Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.698888 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.892850 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.892925 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.893204 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nvds\" (UniqueName: \"kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.893289 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.893347 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.893393 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle\") pod \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\" (UID: \"348d43fe-9095-4a19-b9ac-b3dc9e5da427\") " Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.899546 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.900872 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds" (OuterVolumeSpecName: "kube-api-access-4nvds") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "kube-api-access-4nvds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.901385 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph" (OuterVolumeSpecName: "ceph") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.931236 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory" (OuterVolumeSpecName: "inventory") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.952630 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.952847 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "348d43fe-9095-4a19-b9ac-b3dc9e5da427" (UID: "348d43fe-9095-4a19-b9ac-b3dc9e5da427"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997199 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nvds\" (UniqueName: \"kubernetes.io/projected/348d43fe-9095-4a19-b9ac-b3dc9e5da427-kube-api-access-4nvds\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997258 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997280 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997300 4989 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997319 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:12 crc kubenswrapper[4989]: I1006 10:55:12.997336 4989 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/348d43fe-9095-4a19-b9ac-b3dc9e5da427-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.144964 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-kr879" event={"ID":"348d43fe-9095-4a19-b9ac-b3dc9e5da427","Type":"ContainerDied","Data":"dae2743c3d8ba9334dfd5f9f49733a33b0ab83a6cfc848195919ab06244b659e"} Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.145008 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dae2743c3d8ba9334dfd5f9f49733a33b0ab83a6cfc848195919ab06244b659e" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.145410 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-kr879" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.243533 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-7qfsv"] Oct 06 10:55:13 crc kubenswrapper[4989]: E1006 10:55:13.244230 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="extract-utilities" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244251 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="extract-utilities" Oct 06 10:55:13 crc kubenswrapper[4989]: E1006 10:55:13.244263 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="registry-server" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244271 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="registry-server" Oct 06 10:55:13 crc kubenswrapper[4989]: E1006 10:55:13.244288 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348d43fe-9095-4a19-b9ac-b3dc9e5da427" containerName="libvirt-openstack-openstack-cell1" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244293 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="348d43fe-9095-4a19-b9ac-b3dc9e5da427" containerName="libvirt-openstack-openstack-cell1" Oct 06 10:55:13 crc kubenswrapper[4989]: E1006 10:55:13.244328 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="extract-content" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244333 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="extract-content" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244530 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="348d43fe-9095-4a19-b9ac-b3dc9e5da427" containerName="libvirt-openstack-openstack-cell1" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.244571 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba16934-2d2e-4227-bd91-6c156d8347d1" containerName="registry-server" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.245291 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.248967 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.249300 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.249512 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.249701 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.249956 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.249867 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.250547 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.286945 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-7qfsv"] Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406061 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406237 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406267 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406370 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mzjw\" (UniqueName: \"kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406455 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406505 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406539 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406569 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406589 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.406644 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508513 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508622 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508673 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508696 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508772 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mzjw\" (UniqueName: \"kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508973 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.508991 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.509048 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.509075 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.509096 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.509111 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.511174 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.511777 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.513035 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.513738 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.514294 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.515186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.515612 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.516394 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.519303 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.523231 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.541738 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mzjw\" (UniqueName: \"kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw\") pod \"nova-cell1-openstack-openstack-cell1-7qfsv\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:13 crc kubenswrapper[4989]: I1006 10:55:13.570255 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:55:14 crc kubenswrapper[4989]: I1006 10:55:14.103301 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-7qfsv"] Oct 06 10:55:14 crc kubenswrapper[4989]: I1006 10:55:14.154596 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" event={"ID":"3c116bb3-99b0-4a49-8591-40a42aa83a3d","Type":"ContainerStarted","Data":"1423f5af71ab7ce416a367dc847e714d67300aee9351ac28fe571ff1ffc33eca"} Oct 06 10:55:15 crc kubenswrapper[4989]: I1006 10:55:15.167777 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" event={"ID":"3c116bb3-99b0-4a49-8591-40a42aa83a3d","Type":"ContainerStarted","Data":"66af79868b38897f3232b7b405033d43e4a4df4c914cc3f89e9e55029ba0e12a"} Oct 06 10:55:15 crc kubenswrapper[4989]: I1006 10:55:15.193209 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" podStartSLOduration=2.001506182 podStartE2EDuration="2.193183602s" podCreationTimestamp="2025-10-06 10:55:13 +0000 UTC" firstStartedPulling="2025-10-06 10:55:14.110770358 +0000 UTC m=+8164.900795938" lastFinishedPulling="2025-10-06 10:55:14.302447778 +0000 UTC m=+8165.092473358" observedRunningTime="2025-10-06 10:55:15.187081947 +0000 UTC m=+8165.977107537" watchObservedRunningTime="2025-10-06 10:55:15.193183602 +0000 UTC m=+8165.983209192" Oct 06 10:55:33 crc kubenswrapper[4989]: I1006 10:55:33.935307 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 10:55:33 crc kubenswrapper[4989]: I1006 10:55:33.935853 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 10:55:33 crc kubenswrapper[4989]: I1006 10:55:33.950932 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 10:55:33 crc kubenswrapper[4989]: I1006 10:55:33.952207 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 10:55:33 crc kubenswrapper[4989]: I1006 10:55:33.952273 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" gracePeriod=600 Oct 06 10:55:34 crc kubenswrapper[4989]: E1006 10:55:34.073177 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:55:34 crc kubenswrapper[4989]: I1006 10:55:34.401736 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" exitCode=0 Oct 06 10:55:34 crc kubenswrapper[4989]: I1006 10:55:34.401833 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb"} Oct 06 10:55:34 crc kubenswrapper[4989]: I1006 10:55:34.402119 4989 scope.go:117] "RemoveContainer" containerID="cc61a6c2e1d8fda1b71f0a59f3855607192f5e421389c1e2f0e7c44464718a42" Oct 06 10:55:34 crc kubenswrapper[4989]: I1006 10:55:34.403006 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:55:34 crc kubenswrapper[4989]: E1006 10:55:34.403369 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:55:45 crc kubenswrapper[4989]: I1006 10:55:45.937744 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:55:45 crc kubenswrapper[4989]: E1006 10:55:45.938880 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:55:57 crc kubenswrapper[4989]: I1006 10:55:57.936462 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:55:57 crc kubenswrapper[4989]: E1006 10:55:57.937238 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:56:08 crc kubenswrapper[4989]: I1006 10:56:08.936470 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:56:08 crc kubenswrapper[4989]: E1006 10:56:08.938738 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:56:19 crc kubenswrapper[4989]: I1006 10:56:19.948643 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:56:19 crc kubenswrapper[4989]: E1006 10:56:19.949494 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.683272 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vdqv9"] Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.687206 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.696622 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdqv9"] Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.698836 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-utilities\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.698970 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-catalog-content\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.699026 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zv6b\" (UniqueName: \"kubernetes.io/projected/53b2dca6-6028-4b82-80bc-5b3109054379-kube-api-access-7zv6b\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.799872 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-utilities\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.799978 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-catalog-content\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.800035 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zv6b\" (UniqueName: \"kubernetes.io/projected/53b2dca6-6028-4b82-80bc-5b3109054379-kube-api-access-7zv6b\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.800558 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-utilities\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.800845 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53b2dca6-6028-4b82-80bc-5b3109054379-catalog-content\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:29 crc kubenswrapper[4989]: I1006 10:56:29.822982 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zv6b\" (UniqueName: \"kubernetes.io/projected/53b2dca6-6028-4b82-80bc-5b3109054379-kube-api-access-7zv6b\") pod \"redhat-operators-vdqv9\" (UID: \"53b2dca6-6028-4b82-80bc-5b3109054379\") " pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:30 crc kubenswrapper[4989]: I1006 10:56:30.038940 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:30 crc kubenswrapper[4989]: I1006 10:56:30.538470 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdqv9"] Oct 06 10:56:31 crc kubenswrapper[4989]: I1006 10:56:31.073815 4989 generic.go:334] "Generic (PLEG): container finished" podID="53b2dca6-6028-4b82-80bc-5b3109054379" containerID="07a90f4d69c1c2d13507a859cac98bcbbc676bc31c3ff29e58ea8650bcf60048" exitCode=0 Oct 06 10:56:31 crc kubenswrapper[4989]: I1006 10:56:31.073941 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdqv9" event={"ID":"53b2dca6-6028-4b82-80bc-5b3109054379","Type":"ContainerDied","Data":"07a90f4d69c1c2d13507a859cac98bcbbc676bc31c3ff29e58ea8650bcf60048"} Oct 06 10:56:31 crc kubenswrapper[4989]: I1006 10:56:31.074140 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdqv9" event={"ID":"53b2dca6-6028-4b82-80bc-5b3109054379","Type":"ContainerStarted","Data":"bfd3f2df9fda49b837112374e01491346f8f85966c0fa4f5c51f223b2d99d170"} Oct 06 10:56:31 crc kubenswrapper[4989]: I1006 10:56:31.077401 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 10:56:32 crc kubenswrapper[4989]: I1006 10:56:32.935818 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:56:32 crc kubenswrapper[4989]: E1006 10:56:32.936969 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:56:41 crc kubenswrapper[4989]: I1006 10:56:41.198707 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdqv9" event={"ID":"53b2dca6-6028-4b82-80bc-5b3109054379","Type":"ContainerStarted","Data":"d93a6d7b0deded98c404d457fdaf5249aa9df00bbfc3c53a0f3a3ef4c5cdba3e"} Oct 06 10:56:42 crc kubenswrapper[4989]: I1006 10:56:42.209135 4989 generic.go:334] "Generic (PLEG): container finished" podID="53b2dca6-6028-4b82-80bc-5b3109054379" containerID="d93a6d7b0deded98c404d457fdaf5249aa9df00bbfc3c53a0f3a3ef4c5cdba3e" exitCode=0 Oct 06 10:56:42 crc kubenswrapper[4989]: I1006 10:56:42.209230 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdqv9" event={"ID":"53b2dca6-6028-4b82-80bc-5b3109054379","Type":"ContainerDied","Data":"d93a6d7b0deded98c404d457fdaf5249aa9df00bbfc3c53a0f3a3ef4c5cdba3e"} Oct 06 10:56:43 crc kubenswrapper[4989]: I1006 10:56:43.223080 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdqv9" event={"ID":"53b2dca6-6028-4b82-80bc-5b3109054379","Type":"ContainerStarted","Data":"7d4d56d6517589405342895eb6dffedc14ff5f847b0ab58a84978d2fcd5af22f"} Oct 06 10:56:43 crc kubenswrapper[4989]: I1006 10:56:43.246072 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vdqv9" podStartSLOduration=2.486213962 podStartE2EDuration="14.246050968s" podCreationTimestamp="2025-10-06 10:56:29 +0000 UTC" firstStartedPulling="2025-10-06 10:56:31.077113629 +0000 UTC m=+8241.867139229" lastFinishedPulling="2025-10-06 10:56:42.836950645 +0000 UTC m=+8253.626976235" observedRunningTime="2025-10-06 10:56:43.239132879 +0000 UTC m=+8254.029158459" watchObservedRunningTime="2025-10-06 10:56:43.246050968 +0000 UTC m=+8254.036076548" Oct 06 10:56:43 crc kubenswrapper[4989]: I1006 10:56:43.936598 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:56:43 crc kubenswrapper[4989]: E1006 10:56:43.937029 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.039365 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.041108 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.097382 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.393856 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vdqv9" Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.466542 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdqv9"] Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.512531 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 10:56:50 crc kubenswrapper[4989]: I1006 10:56:50.512799 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-85sl7" podUID="796add64-e739-490b-862c-274297646887" containerName="registry-server" containerID="cri-o://85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd" gracePeriod=2 Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.072438 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.181531 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content\") pod \"796add64-e739-490b-862c-274297646887\" (UID: \"796add64-e739-490b-862c-274297646887\") " Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.181684 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nmpg\" (UniqueName: \"kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg\") pod \"796add64-e739-490b-862c-274297646887\" (UID: \"796add64-e739-490b-862c-274297646887\") " Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.181891 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities\") pod \"796add64-e739-490b-862c-274297646887\" (UID: \"796add64-e739-490b-862c-274297646887\") " Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.182849 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities" (OuterVolumeSpecName: "utilities") pod "796add64-e739-490b-862c-274297646887" (UID: "796add64-e739-490b-862c-274297646887"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.193560 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg" (OuterVolumeSpecName: "kube-api-access-5nmpg") pod "796add64-e739-490b-862c-274297646887" (UID: "796add64-e739-490b-862c-274297646887"). InnerVolumeSpecName "kube-api-access-5nmpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.259022 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "796add64-e739-490b-862c-274297646887" (UID: "796add64-e739-490b-862c-274297646887"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.284431 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.284482 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/796add64-e739-490b-862c-274297646887-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.284499 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nmpg\" (UniqueName: \"kubernetes.io/projected/796add64-e739-490b-862c-274297646887-kube-api-access-5nmpg\") on node \"crc\" DevicePath \"\"" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.318890 4989 generic.go:334] "Generic (PLEG): container finished" podID="796add64-e739-490b-862c-274297646887" containerID="85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd" exitCode=0 Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.319981 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-85sl7" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.324292 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerDied","Data":"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd"} Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.324364 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-85sl7" event={"ID":"796add64-e739-490b-862c-274297646887","Type":"ContainerDied","Data":"98dc2942d603462b5f61bf909447d59c7f2204a6c9c305ee15db4b473285ff17"} Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.325318 4989 scope.go:117] "RemoveContainer" containerID="85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.353494 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.354218 4989 scope.go:117] "RemoveContainer" containerID="dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.361916 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-85sl7"] Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.378004 4989 scope.go:117] "RemoveContainer" containerID="41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.429252 4989 scope.go:117] "RemoveContainer" containerID="85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd" Oct 06 10:56:51 crc kubenswrapper[4989]: E1006 10:56:51.429638 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd\": container with ID starting with 85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd not found: ID does not exist" containerID="85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.429691 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd"} err="failed to get container status \"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd\": rpc error: code = NotFound desc = could not find container \"85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd\": container with ID starting with 85295ef8b1a94682a5a22d4423c6c2adde074e84ac5f421b1e62fe35e9e8fffd not found: ID does not exist" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.429716 4989 scope.go:117] "RemoveContainer" containerID="dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae" Oct 06 10:56:51 crc kubenswrapper[4989]: E1006 10:56:51.430211 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae\": container with ID starting with dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae not found: ID does not exist" containerID="dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.430232 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae"} err="failed to get container status \"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae\": rpc error: code = NotFound desc = could not find container \"dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae\": container with ID starting with dc08472f6088a5c9511931e97a10703457ff999da8df763d0e174a28894b6dae not found: ID does not exist" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.430246 4989 scope.go:117] "RemoveContainer" containerID="41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495" Oct 06 10:56:51 crc kubenswrapper[4989]: E1006 10:56:51.430561 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495\": container with ID starting with 41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495 not found: ID does not exist" containerID="41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.430579 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495"} err="failed to get container status \"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495\": rpc error: code = NotFound desc = could not find container \"41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495\": container with ID starting with 41b8537317d93daa5d4d34f9478cd01677289a42ad75d2c27572ca5826461495 not found: ID does not exist" Oct 06 10:56:51 crc kubenswrapper[4989]: I1006 10:56:51.949415 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="796add64-e739-490b-862c-274297646887" path="/var/lib/kubelet/pods/796add64-e739-490b-862c-274297646887/volumes" Oct 06 10:56:56 crc kubenswrapper[4989]: I1006 10:56:56.939792 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:56:56 crc kubenswrapper[4989]: E1006 10:56:56.940847 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.103883 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:01 crc kubenswrapper[4989]: E1006 10:57:01.105461 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796add64-e739-490b-862c-274297646887" containerName="extract-utilities" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.105482 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="796add64-e739-490b-862c-274297646887" containerName="extract-utilities" Oct 06 10:57:01 crc kubenswrapper[4989]: E1006 10:57:01.105508 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796add64-e739-490b-862c-274297646887" containerName="extract-content" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.105516 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="796add64-e739-490b-862c-274297646887" containerName="extract-content" Oct 06 10:57:01 crc kubenswrapper[4989]: E1006 10:57:01.105541 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="796add64-e739-490b-862c-274297646887" containerName="registry-server" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.105548 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="796add64-e739-490b-862c-274297646887" containerName="registry-server" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.106340 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="796add64-e739-490b-862c-274297646887" containerName="registry-server" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.110081 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.144307 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.208692 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqrl8\" (UniqueName: \"kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.209054 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.209088 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.311011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqrl8\" (UniqueName: \"kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.311077 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.311116 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.311623 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.311670 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.332370 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqrl8\" (UniqueName: \"kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8\") pod \"certified-operators-lbcq6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.439222 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:01 crc kubenswrapper[4989]: I1006 10:57:01.972549 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:02 crc kubenswrapper[4989]: I1006 10:57:02.442237 4989 generic.go:334] "Generic (PLEG): container finished" podID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerID="34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea" exitCode=0 Oct 06 10:57:02 crc kubenswrapper[4989]: I1006 10:57:02.442335 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerDied","Data":"34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea"} Oct 06 10:57:02 crc kubenswrapper[4989]: I1006 10:57:02.446608 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerStarted","Data":"452796964c2372c3575e6dfc9eea0018f41ba5570146c645528f635c20532dfb"} Oct 06 10:57:03 crc kubenswrapper[4989]: I1006 10:57:03.461523 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerStarted","Data":"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627"} Oct 06 10:57:04 crc kubenswrapper[4989]: I1006 10:57:04.474193 4989 generic.go:334] "Generic (PLEG): container finished" podID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerID="bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627" exitCode=0 Oct 06 10:57:04 crc kubenswrapper[4989]: I1006 10:57:04.474240 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerDied","Data":"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627"} Oct 06 10:57:05 crc kubenswrapper[4989]: I1006 10:57:05.489248 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerStarted","Data":"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879"} Oct 06 10:57:05 crc kubenswrapper[4989]: I1006 10:57:05.513985 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lbcq6" podStartSLOduration=1.974296222 podStartE2EDuration="4.513962236s" podCreationTimestamp="2025-10-06 10:57:01 +0000 UTC" firstStartedPulling="2025-10-06 10:57:02.444224275 +0000 UTC m=+8273.234249875" lastFinishedPulling="2025-10-06 10:57:04.983890299 +0000 UTC m=+8275.773915889" observedRunningTime="2025-10-06 10:57:05.504423581 +0000 UTC m=+8276.294449171" watchObservedRunningTime="2025-10-06 10:57:05.513962236 +0000 UTC m=+8276.303987826" Oct 06 10:57:08 crc kubenswrapper[4989]: I1006 10:57:08.936315 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:57:08 crc kubenswrapper[4989]: E1006 10:57:08.937936 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:57:11 crc kubenswrapper[4989]: I1006 10:57:11.440503 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:11 crc kubenswrapper[4989]: I1006 10:57:11.441044 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:11 crc kubenswrapper[4989]: I1006 10:57:11.525690 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:11 crc kubenswrapper[4989]: I1006 10:57:11.659827 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:11 crc kubenswrapper[4989]: I1006 10:57:11.782176 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:13 crc kubenswrapper[4989]: I1006 10:57:13.617385 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lbcq6" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="registry-server" containerID="cri-o://c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879" gracePeriod=2 Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.599586 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.634833 4989 generic.go:334] "Generic (PLEG): container finished" podID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerID="c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879" exitCode=0 Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.634893 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerDied","Data":"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879"} Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.634930 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbcq6" event={"ID":"dab41db7-c4c8-446a-8e6a-a9d67307efe6","Type":"ContainerDied","Data":"452796964c2372c3575e6dfc9eea0018f41ba5570146c645528f635c20532dfb"} Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.634957 4989 scope.go:117] "RemoveContainer" containerID="c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.635154 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbcq6" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.661848 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities\") pod \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.661987 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqrl8\" (UniqueName: \"kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8\") pod \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.662059 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content\") pod \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\" (UID: \"dab41db7-c4c8-446a-8e6a-a9d67307efe6\") " Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.663592 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities" (OuterVolumeSpecName: "utilities") pod "dab41db7-c4c8-446a-8e6a-a9d67307efe6" (UID: "dab41db7-c4c8-446a-8e6a-a9d67307efe6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.670015 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8" (OuterVolumeSpecName: "kube-api-access-hqrl8") pod "dab41db7-c4c8-446a-8e6a-a9d67307efe6" (UID: "dab41db7-c4c8-446a-8e6a-a9d67307efe6"). InnerVolumeSpecName "kube-api-access-hqrl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.676212 4989 scope.go:117] "RemoveContainer" containerID="bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.727153 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dab41db7-c4c8-446a-8e6a-a9d67307efe6" (UID: "dab41db7-c4c8-446a-8e6a-a9d67307efe6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.735419 4989 scope.go:117] "RemoveContainer" containerID="34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.765234 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.765270 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqrl8\" (UniqueName: \"kubernetes.io/projected/dab41db7-c4c8-446a-8e6a-a9d67307efe6-kube-api-access-hqrl8\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.765286 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dab41db7-c4c8-446a-8e6a-a9d67307efe6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.782952 4989 scope.go:117] "RemoveContainer" containerID="c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879" Oct 06 10:57:14 crc kubenswrapper[4989]: E1006 10:57:14.783450 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879\": container with ID starting with c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879 not found: ID does not exist" containerID="c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.783492 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879"} err="failed to get container status \"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879\": rpc error: code = NotFound desc = could not find container \"c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879\": container with ID starting with c6d405367b13f81e3d1d934724e16a2b8cb23ae4608d65868f7c91216ecb5879 not found: ID does not exist" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.783518 4989 scope.go:117] "RemoveContainer" containerID="bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627" Oct 06 10:57:14 crc kubenswrapper[4989]: E1006 10:57:14.784041 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627\": container with ID starting with bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627 not found: ID does not exist" containerID="bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.784067 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627"} err="failed to get container status \"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627\": rpc error: code = NotFound desc = could not find container \"bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627\": container with ID starting with bc128b18bd9bf36c8733d7fd12e3febf2f47e6e241790b0d0ef0b4f3c63c6627 not found: ID does not exist" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.784086 4989 scope.go:117] "RemoveContainer" containerID="34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea" Oct 06 10:57:14 crc kubenswrapper[4989]: E1006 10:57:14.784437 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea\": container with ID starting with 34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea not found: ID does not exist" containerID="34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.784461 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea"} err="failed to get container status \"34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea\": rpc error: code = NotFound desc = could not find container \"34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea\": container with ID starting with 34e6d81ebf111f4e73e1af0c5f1a0a45cc81413c6a4a4f5f29999ac30a23b3ea not found: ID does not exist" Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.976326 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:14 crc kubenswrapper[4989]: I1006 10:57:14.988744 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lbcq6"] Oct 06 10:57:15 crc kubenswrapper[4989]: I1006 10:57:15.974698 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" path="/var/lib/kubelet/pods/dab41db7-c4c8-446a-8e6a-a9d67307efe6/volumes" Oct 06 10:57:19 crc kubenswrapper[4989]: I1006 10:57:19.936870 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:57:19 crc kubenswrapper[4989]: E1006 10:57:19.937438 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:57:31 crc kubenswrapper[4989]: I1006 10:57:31.936900 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:57:31 crc kubenswrapper[4989]: E1006 10:57:31.937689 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:57:44 crc kubenswrapper[4989]: I1006 10:57:44.938045 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:57:44 crc kubenswrapper[4989]: E1006 10:57:44.939511 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:57:57 crc kubenswrapper[4989]: I1006 10:57:57.936789 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:57:57 crc kubenswrapper[4989]: E1006 10:57:57.937704 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:58:12 crc kubenswrapper[4989]: I1006 10:58:12.937029 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:58:12 crc kubenswrapper[4989]: E1006 10:58:12.938264 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:58:24 crc kubenswrapper[4989]: I1006 10:58:24.937969 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:58:24 crc kubenswrapper[4989]: E1006 10:58:24.938939 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:58:37 crc kubenswrapper[4989]: I1006 10:58:37.936511 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:58:37 crc kubenswrapper[4989]: E1006 10:58:37.937594 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:58:51 crc kubenswrapper[4989]: I1006 10:58:51.936262 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:58:51 crc kubenswrapper[4989]: E1006 10:58:51.937242 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:58:54 crc kubenswrapper[4989]: I1006 10:58:54.773905 4989 generic.go:334] "Generic (PLEG): container finished" podID="3c116bb3-99b0-4a49-8591-40a42aa83a3d" containerID="66af79868b38897f3232b7b405033d43e4a4df4c914cc3f89e9e55029ba0e12a" exitCode=0 Oct 06 10:58:54 crc kubenswrapper[4989]: I1006 10:58:54.774054 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" event={"ID":"3c116bb3-99b0-4a49-8591-40a42aa83a3d","Type":"ContainerDied","Data":"66af79868b38897f3232b7b405033d43e4a4df4c914cc3f89e9e55029ba0e12a"} Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.344308 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.453806 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456220 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456342 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456368 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456394 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456622 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456761 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mzjw\" (UniqueName: \"kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.456806 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.457036 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.457169 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key\") pod \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\" (UID: \"3c116bb3-99b0-4a49-8591-40a42aa83a3d\") " Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.470246 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.473759 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph" (OuterVolumeSpecName: "ceph") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.481224 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw" (OuterVolumeSpecName: "kube-api-access-2mzjw") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "kube-api-access-2mzjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.498913 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.503927 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.509893 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.511338 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.513145 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.518563 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.525692 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.534353 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory" (OuterVolumeSpecName: "inventory") pod "3c116bb3-99b0-4a49-8591-40a42aa83a3d" (UID: "3c116bb3-99b0-4a49-8591-40a42aa83a3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560221 4989 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560264 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560277 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560289 4989 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560302 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560318 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560329 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mzjw\" (UniqueName: \"kubernetes.io/projected/3c116bb3-99b0-4a49-8591-40a42aa83a3d-kube-api-access-2mzjw\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560340 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560352 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560363 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.560375 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c116bb3-99b0-4a49-8591-40a42aa83a3d-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.800136 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" event={"ID":"3c116bb3-99b0-4a49-8591-40a42aa83a3d","Type":"ContainerDied","Data":"1423f5af71ab7ce416a367dc847e714d67300aee9351ac28fe571ff1ffc33eca"} Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.800188 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1423f5af71ab7ce416a367dc847e714d67300aee9351ac28fe571ff1ffc33eca" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.800209 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-7qfsv" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.939722 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-4xbqs"] Oct 06 10:58:56 crc kubenswrapper[4989]: E1006 10:58:56.940424 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="extract-utilities" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.940499 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="extract-utilities" Oct 06 10:58:56 crc kubenswrapper[4989]: E1006 10:58:56.940577 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="extract-content" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.940632 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="extract-content" Oct 06 10:58:56 crc kubenswrapper[4989]: E1006 10:58:56.941828 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c116bb3-99b0-4a49-8591-40a42aa83a3d" containerName="nova-cell1-openstack-openstack-cell1" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.941913 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c116bb3-99b0-4a49-8591-40a42aa83a3d" containerName="nova-cell1-openstack-openstack-cell1" Oct 06 10:58:56 crc kubenswrapper[4989]: E1006 10:58:56.941990 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="registry-server" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.942044 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="registry-server" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.942342 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c116bb3-99b0-4a49-8591-40a42aa83a3d" containerName="nova-cell1-openstack-openstack-cell1" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.942453 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab41db7-c4c8-446a-8e6a-a9d67307efe6" containerName="registry-server" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.943948 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.956401 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.956642 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.956956 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.957095 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.957308 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 10:58:56 crc kubenswrapper[4989]: I1006 10:58:56.960854 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-4xbqs"] Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070093 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070288 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070338 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9szr\" (UniqueName: \"kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070427 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070477 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070513 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070540 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.070583 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.172804 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.173258 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.173475 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9szr\" (UniqueName: \"kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.173737 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.173862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.173920 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.174074 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.174201 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.177932 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.178286 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.180031 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.180259 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.180302 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.181003 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.184147 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.196755 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9szr\" (UniqueName: \"kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr\") pod \"telemetry-openstack-openstack-cell1-4xbqs\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.260894 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 10:58:57 crc kubenswrapper[4989]: I1006 10:58:57.858157 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-4xbqs"] Oct 06 10:58:58 crc kubenswrapper[4989]: I1006 10:58:58.828319 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" event={"ID":"68affe53-42a8-400e-824b-af80c5c9e0e4","Type":"ContainerStarted","Data":"a67a326e7b4befa5b6a0d0cfd2b07618051929039d3917a566b980094b22127a"} Oct 06 10:58:58 crc kubenswrapper[4989]: I1006 10:58:58.828931 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" event={"ID":"68affe53-42a8-400e-824b-af80c5c9e0e4","Type":"ContainerStarted","Data":"9d5b384f39f96da29a2055c1040e5aeccc8690939e587dbfa6c74c6304b718c9"} Oct 06 10:58:58 crc kubenswrapper[4989]: I1006 10:58:58.864390 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" podStartSLOduration=2.672090329 podStartE2EDuration="2.864360386s" podCreationTimestamp="2025-10-06 10:58:56 +0000 UTC" firstStartedPulling="2025-10-06 10:58:57.878036309 +0000 UTC m=+8388.668061899" lastFinishedPulling="2025-10-06 10:58:58.070306376 +0000 UTC m=+8388.860331956" observedRunningTime="2025-10-06 10:58:58.852406802 +0000 UTC m=+8389.642432422" watchObservedRunningTime="2025-10-06 10:58:58.864360386 +0000 UTC m=+8389.654386006" Oct 06 10:59:05 crc kubenswrapper[4989]: I1006 10:59:05.937994 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:59:05 crc kubenswrapper[4989]: E1006 10:59:05.939210 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:59:18 crc kubenswrapper[4989]: I1006 10:59:18.935768 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:59:18 crc kubenswrapper[4989]: E1006 10:59:18.936843 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:59:29 crc kubenswrapper[4989]: I1006 10:59:29.944466 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:59:29 crc kubenswrapper[4989]: E1006 10:59:29.946152 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:59:42 crc kubenswrapper[4989]: I1006 10:59:42.937206 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:59:42 crc kubenswrapper[4989]: E1006 10:59:42.937966 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 10:59:55 crc kubenswrapper[4989]: I1006 10:59:55.936262 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 10:59:55 crc kubenswrapper[4989]: E1006 10:59:55.937180 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.152406 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87"] Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.154986 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.157252 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.157852 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.165540 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87"] Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.295585 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f6bz\" (UniqueName: \"kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.296257 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.296433 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.398925 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f6bz\" (UniqueName: \"kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.399019 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.399115 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.400350 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.407985 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.417471 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f6bz\" (UniqueName: \"kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz\") pod \"collect-profiles-29329140-jxn87\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:00 crc kubenswrapper[4989]: I1006 11:00:00.491297 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:01 crc kubenswrapper[4989]: I1006 11:00:01.007009 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87"] Oct 06 11:00:01 crc kubenswrapper[4989]: I1006 11:00:01.548852 4989 generic.go:334] "Generic (PLEG): container finished" podID="a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" containerID="4bb6f7c5c70f98240da9324b6200a17c6a61f5b18070298c7b539ba1d5977b32" exitCode=0 Oct 06 11:00:01 crc kubenswrapper[4989]: I1006 11:00:01.549947 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" event={"ID":"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7","Type":"ContainerDied","Data":"4bb6f7c5c70f98240da9324b6200a17c6a61f5b18070298c7b539ba1d5977b32"} Oct 06 11:00:01 crc kubenswrapper[4989]: I1006 11:00:01.550040 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" event={"ID":"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7","Type":"ContainerStarted","Data":"479f71e7a2062ec7b88b07f22445f846403d36f05f5f7f3a0d581a8c44c07087"} Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.940303 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.959715 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume\") pod \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.959961 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f6bz\" (UniqueName: \"kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz\") pod \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.960559 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume" (OuterVolumeSpecName: "config-volume") pod "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" (UID: "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.961415 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume\") pod \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\" (UID: \"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7\") " Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.962599 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.966018 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" (UID: "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:00:02 crc kubenswrapper[4989]: I1006 11:00:02.967221 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz" (OuterVolumeSpecName: "kube-api-access-4f6bz") pod "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" (UID: "a9bd4c29-7146-4bac-a0f8-34c397f3ffa7"). InnerVolumeSpecName "kube-api-access-4f6bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:00:03 crc kubenswrapper[4989]: I1006 11:00:03.063675 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f6bz\" (UniqueName: \"kubernetes.io/projected/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-kube-api-access-4f6bz\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:03 crc kubenswrapper[4989]: I1006 11:00:03.063876 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a9bd4c29-7146-4bac-a0f8-34c397f3ffa7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:00:03 crc kubenswrapper[4989]: I1006 11:00:03.573640 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" event={"ID":"a9bd4c29-7146-4bac-a0f8-34c397f3ffa7","Type":"ContainerDied","Data":"479f71e7a2062ec7b88b07f22445f846403d36f05f5f7f3a0d581a8c44c07087"} Oct 06 11:00:03 crc kubenswrapper[4989]: I1006 11:00:03.573777 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="479f71e7a2062ec7b88b07f22445f846403d36f05f5f7f3a0d581a8c44c07087" Oct 06 11:00:03 crc kubenswrapper[4989]: I1006 11:00:03.573849 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329140-jxn87" Oct 06 11:00:04 crc kubenswrapper[4989]: I1006 11:00:04.033858 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx"] Oct 06 11:00:04 crc kubenswrapper[4989]: I1006 11:00:04.043247 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329095-xf7gx"] Oct 06 11:00:05 crc kubenswrapper[4989]: I1006 11:00:05.959900 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c6e7a4-1259-43c2-bd73-e7daa82f34b5" path="/var/lib/kubelet/pods/b8c6e7a4-1259-43c2-bd73-e7daa82f34b5/volumes" Oct 06 11:00:08 crc kubenswrapper[4989]: I1006 11:00:08.936705 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 11:00:08 crc kubenswrapper[4989]: E1006 11:00:08.937452 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:00:19 crc kubenswrapper[4989]: I1006 11:00:19.952204 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 11:00:19 crc kubenswrapper[4989]: E1006 11:00:19.953413 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:00:27 crc kubenswrapper[4989]: I1006 11:00:27.622845 4989 scope.go:117] "RemoveContainer" containerID="6b2b51b00ab758b5d651e1645075b3ed8c0bf450616374096fa7635dc4f4ca57" Oct 06 11:00:30 crc kubenswrapper[4989]: I1006 11:00:30.936621 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 11:00:30 crc kubenswrapper[4989]: E1006 11:00:30.937772 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:00:44 crc kubenswrapper[4989]: I1006 11:00:44.936228 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 11:00:46 crc kubenswrapper[4989]: I1006 11:00:46.044846 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f"} Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.171430 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29329141-9vw8m"] Oct 06 11:01:00 crc kubenswrapper[4989]: E1006 11:01:00.172528 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" containerName="collect-profiles" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.172546 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" containerName="collect-profiles" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.172960 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9bd4c29-7146-4bac-a0f8-34c397f3ffa7" containerName="collect-profiles" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.173911 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.188569 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29329141-9vw8m"] Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.335396 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.335805 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.335844 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx4q2\" (UniqueName: \"kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.335886 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.437515 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.437623 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.437727 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.437773 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx4q2\" (UniqueName: \"kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.446052 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.446965 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.447137 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.454825 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx4q2\" (UniqueName: \"kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2\") pod \"keystone-cron-29329141-9vw8m\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.505192 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:00 crc kubenswrapper[4989]: I1006 11:01:00.972325 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29329141-9vw8m"] Oct 06 11:01:01 crc kubenswrapper[4989]: I1006 11:01:01.211294 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-9vw8m" event={"ID":"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a","Type":"ContainerStarted","Data":"05894647c00bf88592c841699c72e78c8540c3ec2a06d5589796aec7c90ac3f8"} Oct 06 11:01:01 crc kubenswrapper[4989]: I1006 11:01:01.211694 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-9vw8m" event={"ID":"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a","Type":"ContainerStarted","Data":"a839a24a9f5d8bc1b12c7acfe0f72046435e04c056a91bd356b7fd0562763758"} Oct 06 11:01:01 crc kubenswrapper[4989]: I1006 11:01:01.231801 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29329141-9vw8m" podStartSLOduration=1.231781499 podStartE2EDuration="1.231781499s" podCreationTimestamp="2025-10-06 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:01:01.228329599 +0000 UTC m=+8512.018355189" watchObservedRunningTime="2025-10-06 11:01:01.231781499 +0000 UTC m=+8512.021807099" Oct 06 11:01:04 crc kubenswrapper[4989]: I1006 11:01:04.248252 4989 generic.go:334] "Generic (PLEG): container finished" podID="a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" containerID="05894647c00bf88592c841699c72e78c8540c3ec2a06d5589796aec7c90ac3f8" exitCode=0 Oct 06 11:01:04 crc kubenswrapper[4989]: I1006 11:01:04.248338 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-9vw8m" event={"ID":"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a","Type":"ContainerDied","Data":"05894647c00bf88592c841699c72e78c8540c3ec2a06d5589796aec7c90ac3f8"} Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.591173 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.758195 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data\") pod \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.758241 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle\") pod \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.758316 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys\") pod \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.758456 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx4q2\" (UniqueName: \"kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2\") pod \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\" (UID: \"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a\") " Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.763353 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2" (OuterVolumeSpecName: "kube-api-access-lx4q2") pod "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" (UID: "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a"). InnerVolumeSpecName "kube-api-access-lx4q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.763908 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" (UID: "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.804257 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" (UID: "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.830154 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data" (OuterVolumeSpecName: "config-data") pod "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" (UID: "a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.861006 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx4q2\" (UniqueName: \"kubernetes.io/projected/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-kube-api-access-lx4q2\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.861035 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.861044 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:05 crc kubenswrapper[4989]: I1006 11:01:05.861053 4989 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 06 11:01:06 crc kubenswrapper[4989]: I1006 11:01:06.268880 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29329141-9vw8m" event={"ID":"a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a","Type":"ContainerDied","Data":"a839a24a9f5d8bc1b12c7acfe0f72046435e04c056a91bd356b7fd0562763758"} Oct 06 11:01:06 crc kubenswrapper[4989]: I1006 11:01:06.268923 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a839a24a9f5d8bc1b12c7acfe0f72046435e04c056a91bd356b7fd0562763758" Oct 06 11:01:06 crc kubenswrapper[4989]: I1006 11:01:06.268949 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29329141-9vw8m" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.687893 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:19 crc kubenswrapper[4989]: E1006 11:02:19.689012 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" containerName="keystone-cron" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.689029 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" containerName="keystone-cron" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.689361 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a" containerName="keystone-cron" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.691177 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.705841 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.722807 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.722939 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj97\" (UniqueName: \"kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.723070 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.825303 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.825453 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.825540 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj97\" (UniqueName: \"kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.825889 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.825918 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:19 crc kubenswrapper[4989]: I1006 11:02:19.847077 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj97\" (UniqueName: \"kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97\") pod \"redhat-marketplace-rhlwj\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:20 crc kubenswrapper[4989]: I1006 11:02:20.022384 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:20 crc kubenswrapper[4989]: I1006 11:02:20.489988 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:21 crc kubenswrapper[4989]: I1006 11:02:21.124518 4989 generic.go:334] "Generic (PLEG): container finished" podID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerID="dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52" exitCode=0 Oct 06 11:02:21 crc kubenswrapper[4989]: I1006 11:02:21.124813 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerDied","Data":"dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52"} Oct 06 11:02:21 crc kubenswrapper[4989]: I1006 11:02:21.125111 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerStarted","Data":"9df1921c35b648679d7ed0925dcff8520e36dde55d929b3431364aaebf356888"} Oct 06 11:02:21 crc kubenswrapper[4989]: I1006 11:02:21.128969 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:02:22 crc kubenswrapper[4989]: I1006 11:02:22.139466 4989 generic.go:334] "Generic (PLEG): container finished" podID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerID="6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe" exitCode=0 Oct 06 11:02:22 crc kubenswrapper[4989]: I1006 11:02:22.139539 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerDied","Data":"6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe"} Oct 06 11:02:23 crc kubenswrapper[4989]: I1006 11:02:23.150983 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerStarted","Data":"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667"} Oct 06 11:02:23 crc kubenswrapper[4989]: I1006 11:02:23.178553 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rhlwj" podStartSLOduration=2.766644445 podStartE2EDuration="4.178536869s" podCreationTimestamp="2025-10-06 11:02:19 +0000 UTC" firstStartedPulling="2025-10-06 11:02:21.128602029 +0000 UTC m=+8591.918627609" lastFinishedPulling="2025-10-06 11:02:22.540494453 +0000 UTC m=+8593.330520033" observedRunningTime="2025-10-06 11:02:23.168272424 +0000 UTC m=+8593.958298034" watchObservedRunningTime="2025-10-06 11:02:23.178536869 +0000 UTC m=+8593.968562449" Oct 06 11:02:30 crc kubenswrapper[4989]: I1006 11:02:30.022547 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:30 crc kubenswrapper[4989]: I1006 11:02:30.023189 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:30 crc kubenswrapper[4989]: I1006 11:02:30.081884 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:30 crc kubenswrapper[4989]: I1006 11:02:30.269788 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:30 crc kubenswrapper[4989]: I1006 11:02:30.322296 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.245041 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rhlwj" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="registry-server" containerID="cri-o://6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667" gracePeriod=2 Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.813434 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.924505 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content\") pod \"68ffd3b0-47e0-441e-8b27-39e512e42a15\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.924606 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities\") pod \"68ffd3b0-47e0-441e-8b27-39e512e42a15\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.925518 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities" (OuterVolumeSpecName: "utilities") pod "68ffd3b0-47e0-441e-8b27-39e512e42a15" (UID: "68ffd3b0-47e0-441e-8b27-39e512e42a15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.926340 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsj97\" (UniqueName: \"kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97\") pod \"68ffd3b0-47e0-441e-8b27-39e512e42a15\" (UID: \"68ffd3b0-47e0-441e-8b27-39e512e42a15\") " Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.926954 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.937358 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97" (OuterVolumeSpecName: "kube-api-access-qsj97") pod "68ffd3b0-47e0-441e-8b27-39e512e42a15" (UID: "68ffd3b0-47e0-441e-8b27-39e512e42a15"). InnerVolumeSpecName "kube-api-access-qsj97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:02:32 crc kubenswrapper[4989]: I1006 11:02:32.950096 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68ffd3b0-47e0-441e-8b27-39e512e42a15" (UID: "68ffd3b0-47e0-441e-8b27-39e512e42a15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.029017 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ffd3b0-47e0-441e-8b27-39e512e42a15-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.029061 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsj97\" (UniqueName: \"kubernetes.io/projected/68ffd3b0-47e0-441e-8b27-39e512e42a15-kube-api-access-qsj97\") on node \"crc\" DevicePath \"\"" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.256706 4989 generic.go:334] "Generic (PLEG): container finished" podID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerID="6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667" exitCode=0 Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.256783 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerDied","Data":"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667"} Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.256822 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rhlwj" event={"ID":"68ffd3b0-47e0-441e-8b27-39e512e42a15","Type":"ContainerDied","Data":"9df1921c35b648679d7ed0925dcff8520e36dde55d929b3431364aaebf356888"} Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.256840 4989 scope.go:117] "RemoveContainer" containerID="6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.256790 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rhlwj" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.290269 4989 scope.go:117] "RemoveContainer" containerID="6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.296533 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.305702 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rhlwj"] Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.313197 4989 scope.go:117] "RemoveContainer" containerID="dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.363040 4989 scope.go:117] "RemoveContainer" containerID="6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667" Oct 06 11:02:33 crc kubenswrapper[4989]: E1006 11:02:33.363486 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667\": container with ID starting with 6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667 not found: ID does not exist" containerID="6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.363527 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667"} err="failed to get container status \"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667\": rpc error: code = NotFound desc = could not find container \"6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667\": container with ID starting with 6ca9672a299cb1b5170ce523dea2439fdc6b205e25533ec69e3c347a6028f667 not found: ID does not exist" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.363554 4989 scope.go:117] "RemoveContainer" containerID="6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe" Oct 06 11:02:33 crc kubenswrapper[4989]: E1006 11:02:33.363966 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe\": container with ID starting with 6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe not found: ID does not exist" containerID="6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.363995 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe"} err="failed to get container status \"6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe\": rpc error: code = NotFound desc = could not find container \"6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe\": container with ID starting with 6bfe4f3ee2ffb1c10f26ce71b766372cc3cb438454f138400bfd136ce8f5ccfe not found: ID does not exist" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.364015 4989 scope.go:117] "RemoveContainer" containerID="dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52" Oct 06 11:02:33 crc kubenswrapper[4989]: E1006 11:02:33.364454 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52\": container with ID starting with dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52 not found: ID does not exist" containerID="dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.364492 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52"} err="failed to get container status \"dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52\": rpc error: code = NotFound desc = could not find container \"dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52\": container with ID starting with dfd96441e480775f1036bc8aed36d733dd6346b887e7febddc8241503c0b1f52 not found: ID does not exist" Oct 06 11:02:33 crc kubenswrapper[4989]: I1006 11:02:33.947743 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" path="/var/lib/kubelet/pods/68ffd3b0-47e0-441e-8b27-39e512e42a15/volumes" Oct 06 11:03:03 crc kubenswrapper[4989]: I1006 11:03:03.935400 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:03:03 crc kubenswrapper[4989]: I1006 11:03:03.936021 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:03:29 crc kubenswrapper[4989]: I1006 11:03:29.940349 4989 generic.go:334] "Generic (PLEG): container finished" podID="68affe53-42a8-400e-824b-af80c5c9e0e4" containerID="a67a326e7b4befa5b6a0d0cfd2b07618051929039d3917a566b980094b22127a" exitCode=0 Oct 06 11:03:29 crc kubenswrapper[4989]: I1006 11:03:29.965981 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" event={"ID":"68affe53-42a8-400e-824b-af80c5c9e0e4","Type":"ContainerDied","Data":"a67a326e7b4befa5b6a0d0cfd2b07618051929039d3917a566b980094b22127a"} Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.884477 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:30 crc kubenswrapper[4989]: E1006 11:03:30.885434 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="extract-content" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.885458 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="extract-content" Oct 06 11:03:30 crc kubenswrapper[4989]: E1006 11:03:30.885492 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="extract-utilities" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.885500 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="extract-utilities" Oct 06 11:03:30 crc kubenswrapper[4989]: E1006 11:03:30.885527 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="registry-server" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.885536 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="registry-server" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.885815 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ffd3b0-47e0-441e-8b27-39e512e42a15" containerName="registry-server" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.888305 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.908578 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.962672 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.963089 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:30 crc kubenswrapper[4989]: I1006 11:03:30.963251 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8qwm\" (UniqueName: \"kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.064755 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.064880 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8qwm\" (UniqueName: \"kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.064987 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.065240 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.065366 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.086347 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8qwm\" (UniqueName: \"kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm\") pod \"community-operators-ntk5j\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.217466 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.511335 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687493 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687644 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687720 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687779 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9szr\" (UniqueName: \"kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687810 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687887 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.687933 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.688008 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key\") pod \"68affe53-42a8-400e-824b-af80c5c9e0e4\" (UID: \"68affe53-42a8-400e-824b-af80c5c9e0e4\") " Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.694707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.694814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr" (OuterVolumeSpecName: "kube-api-access-g9szr") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "kube-api-access-g9szr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.695707 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph" (OuterVolumeSpecName: "ceph") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.726581 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.726882 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.731054 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.734916 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.735304 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory" (OuterVolumeSpecName: "inventory") pod "68affe53-42a8-400e-824b-af80c5c9e0e4" (UID: "68affe53-42a8-400e-824b-af80c5c9e0e4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.789997 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790036 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790054 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790101 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790115 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9szr\" (UniqueName: \"kubernetes.io/projected/68affe53-42a8-400e-824b-af80c5c9e0e4-kube-api-access-g9szr\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790127 4989 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790138 4989 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.790150 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/68affe53-42a8-400e-824b-af80c5c9e0e4-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.791399 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.961921 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" event={"ID":"68affe53-42a8-400e-824b-af80c5c9e0e4","Type":"ContainerDied","Data":"9d5b384f39f96da29a2055c1040e5aeccc8690939e587dbfa6c74c6304b718c9"} Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.961960 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d5b384f39f96da29a2055c1040e5aeccc8690939e587dbfa6c74c6304b718c9" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.962009 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-4xbqs" Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.964856 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerStarted","Data":"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3"} Oct 06 11:03:31 crc kubenswrapper[4989]: I1006 11:03:31.964905 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerStarted","Data":"78aa14feea8f910ae6fd57c6cccb4b4f300ddff44c68186c266102769e6a9da9"} Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.082171 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5hmfm"] Oct 06 11:03:32 crc kubenswrapper[4989]: E1006 11:03:32.082711 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68affe53-42a8-400e-824b-af80c5c9e0e4" containerName="telemetry-openstack-openstack-cell1" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.082730 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="68affe53-42a8-400e-824b-af80c5c9e0e4" containerName="telemetry-openstack-openstack-cell1" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.082957 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="68affe53-42a8-400e-824b-af80c5c9e0e4" containerName="telemetry-openstack-openstack-cell1" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.083932 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.090564 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.091360 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.091573 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.092196 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5hmfm"] Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.092372 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.108093 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199011 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199154 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9n8z\" (UniqueName: \"kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199188 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199232 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199267 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.199287 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.301959 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.302343 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.302522 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.302815 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.303353 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9n8z\" (UniqueName: \"kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.303594 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.308343 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.308586 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.310568 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.310713 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.310968 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.321630 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9n8z\" (UniqueName: \"kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z\") pod \"neutron-sriov-openstack-openstack-cell1-5hmfm\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.468804 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.985868 4989 generic.go:334] "Generic (PLEG): container finished" podID="84d8f971-235e-49bc-b3e9-17957bee264d" containerID="afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3" exitCode=0 Oct 06 11:03:32 crc kubenswrapper[4989]: I1006 11:03:32.986214 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerDied","Data":"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3"} Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.081264 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5hmfm"] Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.937582 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.938002 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.996504 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerStarted","Data":"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926"} Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.999425 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" event={"ID":"d24e47c9-fab9-49dc-9fb3-728e9611a342","Type":"ContainerStarted","Data":"e62eba34e7daf6f024da57b89f5304dcc474b33a8497d2c914d1ed0787cf2f7b"} Oct 06 11:03:33 crc kubenswrapper[4989]: I1006 11:03:33.999488 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" event={"ID":"d24e47c9-fab9-49dc-9fb3-728e9611a342","Type":"ContainerStarted","Data":"5086d3104965445ec352f1d87492cd69cbb6f74cbe43272055a7544325951e52"} Oct 06 11:03:34 crc kubenswrapper[4989]: I1006 11:03:34.045061 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" podStartSLOduration=1.893333058 podStartE2EDuration="2.045040237s" podCreationTimestamp="2025-10-06 11:03:32 +0000 UTC" firstStartedPulling="2025-10-06 11:03:33.09031408 +0000 UTC m=+8663.880339660" lastFinishedPulling="2025-10-06 11:03:33.242021219 +0000 UTC m=+8664.032046839" observedRunningTime="2025-10-06 11:03:34.038838829 +0000 UTC m=+8664.828864409" watchObservedRunningTime="2025-10-06 11:03:34.045040237 +0000 UTC m=+8664.835065827" Oct 06 11:03:35 crc kubenswrapper[4989]: I1006 11:03:35.010493 4989 generic.go:334] "Generic (PLEG): container finished" podID="84d8f971-235e-49bc-b3e9-17957bee264d" containerID="8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926" exitCode=0 Oct 06 11:03:35 crc kubenswrapper[4989]: I1006 11:03:35.010580 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerDied","Data":"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926"} Oct 06 11:03:36 crc kubenswrapper[4989]: I1006 11:03:36.025512 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerStarted","Data":"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3"} Oct 06 11:03:36 crc kubenswrapper[4989]: I1006 11:03:36.063678 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntk5j" podStartSLOduration=3.599430833 podStartE2EDuration="6.063641706s" podCreationTimestamp="2025-10-06 11:03:30 +0000 UTC" firstStartedPulling="2025-10-06 11:03:32.989055813 +0000 UTC m=+8663.779081433" lastFinishedPulling="2025-10-06 11:03:35.453266726 +0000 UTC m=+8666.243292306" observedRunningTime="2025-10-06 11:03:36.045543574 +0000 UTC m=+8666.835569174" watchObservedRunningTime="2025-10-06 11:03:36.063641706 +0000 UTC m=+8666.853667286" Oct 06 11:03:41 crc kubenswrapper[4989]: I1006 11:03:41.217623 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:41 crc kubenswrapper[4989]: I1006 11:03:41.218297 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:41 crc kubenswrapper[4989]: I1006 11:03:41.291030 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:42 crc kubenswrapper[4989]: I1006 11:03:42.214245 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:42 crc kubenswrapper[4989]: I1006 11:03:42.297678 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.142203 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ntk5j" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="registry-server" containerID="cri-o://0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3" gracePeriod=2 Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.707387 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.828606 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8qwm\" (UniqueName: \"kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm\") pod \"84d8f971-235e-49bc-b3e9-17957bee264d\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.828909 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content\") pod \"84d8f971-235e-49bc-b3e9-17957bee264d\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.829009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities\") pod \"84d8f971-235e-49bc-b3e9-17957bee264d\" (UID: \"84d8f971-235e-49bc-b3e9-17957bee264d\") " Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.830307 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities" (OuterVolumeSpecName: "utilities") pod "84d8f971-235e-49bc-b3e9-17957bee264d" (UID: "84d8f971-235e-49bc-b3e9-17957bee264d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.834960 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm" (OuterVolumeSpecName: "kube-api-access-h8qwm") pod "84d8f971-235e-49bc-b3e9-17957bee264d" (UID: "84d8f971-235e-49bc-b3e9-17957bee264d"). InnerVolumeSpecName "kube-api-access-h8qwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.879983 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84d8f971-235e-49bc-b3e9-17957bee264d" (UID: "84d8f971-235e-49bc-b3e9-17957bee264d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.931310 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.931351 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d8f971-235e-49bc-b3e9-17957bee264d-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:44 crc kubenswrapper[4989]: I1006 11:03:44.931362 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8qwm\" (UniqueName: \"kubernetes.io/projected/84d8f971-235e-49bc-b3e9-17957bee264d-kube-api-access-h8qwm\") on node \"crc\" DevicePath \"\"" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.155905 4989 generic.go:334] "Generic (PLEG): container finished" podID="84d8f971-235e-49bc-b3e9-17957bee264d" containerID="0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3" exitCode=0 Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.155998 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntk5j" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.156088 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerDied","Data":"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3"} Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.156782 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntk5j" event={"ID":"84d8f971-235e-49bc-b3e9-17957bee264d","Type":"ContainerDied","Data":"78aa14feea8f910ae6fd57c6cccb4b4f300ddff44c68186c266102769e6a9da9"} Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.156808 4989 scope.go:117] "RemoveContainer" containerID="0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.190554 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.194426 4989 scope.go:117] "RemoveContainer" containerID="8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.201483 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ntk5j"] Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.233282 4989 scope.go:117] "RemoveContainer" containerID="afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.290469 4989 scope.go:117] "RemoveContainer" containerID="0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3" Oct 06 11:03:45 crc kubenswrapper[4989]: E1006 11:03:45.291018 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3\": container with ID starting with 0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3 not found: ID does not exist" containerID="0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.291072 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3"} err="failed to get container status \"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3\": rpc error: code = NotFound desc = could not find container \"0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3\": container with ID starting with 0927fcd6f69c29605c7e37ddf668190f239ac945f1aec05e95d412ed7841fec3 not found: ID does not exist" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.291094 4989 scope.go:117] "RemoveContainer" containerID="8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926" Oct 06 11:03:45 crc kubenswrapper[4989]: E1006 11:03:45.291339 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926\": container with ID starting with 8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926 not found: ID does not exist" containerID="8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.291369 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926"} err="failed to get container status \"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926\": rpc error: code = NotFound desc = could not find container \"8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926\": container with ID starting with 8ad7c983dfcfead2e7739a33c0dd20ab709c5d6e877d63930cdd7d9798e4c926 not found: ID does not exist" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.291389 4989 scope.go:117] "RemoveContainer" containerID="afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3" Oct 06 11:03:45 crc kubenswrapper[4989]: E1006 11:03:45.292720 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3\": container with ID starting with afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3 not found: ID does not exist" containerID="afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.292745 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3"} err="failed to get container status \"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3\": rpc error: code = NotFound desc = could not find container \"afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3\": container with ID starting with afa7b4555fde169f142a3050e264ffe2ddf67aa9835360deaa0abf83d5c8abd3 not found: ID does not exist" Oct 06 11:03:45 crc kubenswrapper[4989]: I1006 11:03:45.950590 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" path="/var/lib/kubelet/pods/84d8f971-235e-49bc-b3e9-17957bee264d/volumes" Oct 06 11:04:03 crc kubenswrapper[4989]: I1006 11:04:03.935620 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:04:03 crc kubenswrapper[4989]: I1006 11:04:03.936505 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:04:03 crc kubenswrapper[4989]: I1006 11:04:03.960883 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:04:03 crc kubenswrapper[4989]: I1006 11:04:03.962894 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:04:03 crc kubenswrapper[4989]: I1006 11:04:03.963045 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f" gracePeriod=600 Oct 06 11:04:04 crc kubenswrapper[4989]: I1006 11:04:04.383523 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f" exitCode=0 Oct 06 11:04:04 crc kubenswrapper[4989]: I1006 11:04:04.383571 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f"} Oct 06 11:04:04 crc kubenswrapper[4989]: I1006 11:04:04.383866 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54"} Oct 06 11:04:04 crc kubenswrapper[4989]: I1006 11:04:04.383886 4989 scope.go:117] "RemoveContainer" containerID="7bb90ef7a771f35bcd28b672e97a23c47b88fcd7f6d5c9dfb653083754004bdb" Oct 06 11:06:33 crc kubenswrapper[4989]: I1006 11:06:33.935819 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:06:33 crc kubenswrapper[4989]: I1006 11:06:33.936594 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:07:03 crc kubenswrapper[4989]: I1006 11:07:03.935516 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:07:03 crc kubenswrapper[4989]: I1006 11:07:03.936100 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:07:33 crc kubenswrapper[4989]: I1006 11:07:33.936222 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:07:33 crc kubenswrapper[4989]: I1006 11:07:33.938422 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:07:33 crc kubenswrapper[4989]: I1006 11:07:33.957160 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:07:33 crc kubenswrapper[4989]: I1006 11:07:33.957961 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:07:33 crc kubenswrapper[4989]: I1006 11:07:33.958024 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" gracePeriod=600 Oct 06 11:07:34 crc kubenswrapper[4989]: E1006 11:07:34.082157 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:07:34 crc kubenswrapper[4989]: I1006 11:07:34.919476 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" exitCode=0 Oct 06 11:07:34 crc kubenswrapper[4989]: I1006 11:07:34.919511 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54"} Oct 06 11:07:34 crc kubenswrapper[4989]: I1006 11:07:34.919829 4989 scope.go:117] "RemoveContainer" containerID="b799db9d07b8f6bad02f5113ad6b282bba83102d60c0aabdcad108bf1321627f" Oct 06 11:07:34 crc kubenswrapper[4989]: I1006 11:07:34.920480 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:07:34 crc kubenswrapper[4989]: E1006 11:07:34.920927 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.044561 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:42 crc kubenswrapper[4989]: E1006 11:07:42.046177 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="extract-content" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.046209 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="extract-content" Oct 06 11:07:42 crc kubenswrapper[4989]: E1006 11:07:42.046289 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="extract-utilities" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.046308 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="extract-utilities" Oct 06 11:07:42 crc kubenswrapper[4989]: E1006 11:07:42.046368 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="registry-server" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.046385 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="registry-server" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.046955 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d8f971-235e-49bc-b3e9-17957bee264d" containerName="registry-server" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.051164 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.085091 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.223012 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nph9l\" (UniqueName: \"kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.223075 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.223264 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.325736 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nph9l\" (UniqueName: \"kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.325806 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.326022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.326567 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.326588 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.354892 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nph9l\" (UniqueName: \"kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l\") pod \"certified-operators-nrjdm\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:42 crc kubenswrapper[4989]: I1006 11:07:42.389884 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:43 crc kubenswrapper[4989]: I1006 11:07:43.060865 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:43 crc kubenswrapper[4989]: W1006 11:07:43.067165 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod267db010_74d9_441b_9b41_ae1825fef07e.slice/crio-2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90 WatchSource:0}: Error finding container 2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90: Status 404 returned error can't find the container with id 2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90 Oct 06 11:07:44 crc kubenswrapper[4989]: I1006 11:07:44.044386 4989 generic.go:334] "Generic (PLEG): container finished" podID="267db010-74d9-441b-9b41-ae1825fef07e" containerID="9d8088dc2276669ba6b3bd6460d5c18fcf6559d52cfcd7f4aa47c56a911288de" exitCode=0 Oct 06 11:07:44 crc kubenswrapper[4989]: I1006 11:07:44.044513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerDied","Data":"9d8088dc2276669ba6b3bd6460d5c18fcf6559d52cfcd7f4aa47c56a911288de"} Oct 06 11:07:44 crc kubenswrapper[4989]: I1006 11:07:44.045758 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerStarted","Data":"2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90"} Oct 06 11:07:44 crc kubenswrapper[4989]: I1006 11:07:44.050030 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:07:45 crc kubenswrapper[4989]: I1006 11:07:45.060121 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerStarted","Data":"0d629e2458e0dfb1cc7825ce525812276fa44d6e516f4ec916b1d2385eccc880"} Oct 06 11:07:46 crc kubenswrapper[4989]: I1006 11:07:46.073791 4989 generic.go:334] "Generic (PLEG): container finished" podID="267db010-74d9-441b-9b41-ae1825fef07e" containerID="0d629e2458e0dfb1cc7825ce525812276fa44d6e516f4ec916b1d2385eccc880" exitCode=0 Oct 06 11:07:46 crc kubenswrapper[4989]: I1006 11:07:46.073847 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerDied","Data":"0d629e2458e0dfb1cc7825ce525812276fa44d6e516f4ec916b1d2385eccc880"} Oct 06 11:07:47 crc kubenswrapper[4989]: I1006 11:07:47.092629 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerStarted","Data":"67bc295da46b455fae79eb069817722655b643277f36989f3f11c503b0d0aa12"} Oct 06 11:07:47 crc kubenswrapper[4989]: I1006 11:07:47.123201 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nrjdm" podStartSLOduration=2.53230225 podStartE2EDuration="5.123182531s" podCreationTimestamp="2025-10-06 11:07:42 +0000 UTC" firstStartedPulling="2025-10-06 11:07:44.049756823 +0000 UTC m=+8914.839782403" lastFinishedPulling="2025-10-06 11:07:46.640637094 +0000 UTC m=+8917.430662684" observedRunningTime="2025-10-06 11:07:47.114014907 +0000 UTC m=+8917.904040507" watchObservedRunningTime="2025-10-06 11:07:47.123182531 +0000 UTC m=+8917.913208111" Oct 06 11:07:49 crc kubenswrapper[4989]: I1006 11:07:49.943785 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:07:49 crc kubenswrapper[4989]: E1006 11:07:49.946277 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:07:52 crc kubenswrapper[4989]: I1006 11:07:52.395531 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:52 crc kubenswrapper[4989]: I1006 11:07:52.395927 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:52 crc kubenswrapper[4989]: I1006 11:07:52.464396 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:53 crc kubenswrapper[4989]: I1006 11:07:53.569345 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:53 crc kubenswrapper[4989]: I1006 11:07:53.631818 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:55 crc kubenswrapper[4989]: I1006 11:07:55.187575 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nrjdm" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="registry-server" containerID="cri-o://67bc295da46b455fae79eb069817722655b643277f36989f3f11c503b0d0aa12" gracePeriod=2 Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.210974 4989 generic.go:334] "Generic (PLEG): container finished" podID="267db010-74d9-441b-9b41-ae1825fef07e" containerID="67bc295da46b455fae79eb069817722655b643277f36989f3f11c503b0d0aa12" exitCode=0 Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.211063 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerDied","Data":"67bc295da46b455fae79eb069817722655b643277f36989f3f11c503b0d0aa12"} Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.211353 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrjdm" event={"ID":"267db010-74d9-441b-9b41-ae1825fef07e","Type":"ContainerDied","Data":"2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90"} Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.211374 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2833f326fe9c45212a6a08f6593c07b4ac740a25f1916fedac4db0c21eeccc90" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.232591 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.293184 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nph9l\" (UniqueName: \"kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l\") pod \"267db010-74d9-441b-9b41-ae1825fef07e\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.293251 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content\") pod \"267db010-74d9-441b-9b41-ae1825fef07e\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.293308 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities\") pod \"267db010-74d9-441b-9b41-ae1825fef07e\" (UID: \"267db010-74d9-441b-9b41-ae1825fef07e\") " Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.295059 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities" (OuterVolumeSpecName: "utilities") pod "267db010-74d9-441b-9b41-ae1825fef07e" (UID: "267db010-74d9-441b-9b41-ae1825fef07e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.303608 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l" (OuterVolumeSpecName: "kube-api-access-nph9l") pod "267db010-74d9-441b-9b41-ae1825fef07e" (UID: "267db010-74d9-441b-9b41-ae1825fef07e"). InnerVolumeSpecName "kube-api-access-nph9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.350830 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "267db010-74d9-441b-9b41-ae1825fef07e" (UID: "267db010-74d9-441b-9b41-ae1825fef07e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.397059 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.397120 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nph9l\" (UniqueName: \"kubernetes.io/projected/267db010-74d9-441b-9b41-ae1825fef07e-kube-api-access-nph9l\") on node \"crc\" DevicePath \"\"" Oct 06 11:07:56 crc kubenswrapper[4989]: I1006 11:07:56.397218 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267db010-74d9-441b-9b41-ae1825fef07e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:07:57 crc kubenswrapper[4989]: I1006 11:07:57.220626 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrjdm" Oct 06 11:07:57 crc kubenswrapper[4989]: I1006 11:07:57.275955 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:57 crc kubenswrapper[4989]: I1006 11:07:57.292723 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nrjdm"] Oct 06 11:07:57 crc kubenswrapper[4989]: I1006 11:07:57.963979 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="267db010-74d9-441b-9b41-ae1825fef07e" path="/var/lib/kubelet/pods/267db010-74d9-441b-9b41-ae1825fef07e/volumes" Oct 06 11:08:03 crc kubenswrapper[4989]: I1006 11:08:03.935901 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:08:03 crc kubenswrapper[4989]: E1006 11:08:03.936934 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:08:17 crc kubenswrapper[4989]: I1006 11:08:17.936365 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:08:17 crc kubenswrapper[4989]: E1006 11:08:17.937222 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:08:29 crc kubenswrapper[4989]: I1006 11:08:29.966059 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:08:29 crc kubenswrapper[4989]: E1006 11:08:29.966981 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:08:39 crc kubenswrapper[4989]: I1006 11:08:39.698961 4989 generic.go:334] "Generic (PLEG): container finished" podID="d24e47c9-fab9-49dc-9fb3-728e9611a342" containerID="e62eba34e7daf6f024da57b89f5304dcc474b33a8497d2c914d1ed0787cf2f7b" exitCode=0 Oct 06 11:08:39 crc kubenswrapper[4989]: I1006 11:08:39.698999 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" event={"ID":"d24e47c9-fab9-49dc-9fb3-728e9611a342","Type":"ContainerDied","Data":"e62eba34e7daf6f024da57b89f5304dcc474b33a8497d2c914d1ed0787cf2f7b"} Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.167839 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.243767 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.244139 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.244310 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.244460 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.244501 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.244599 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9n8z\" (UniqueName: \"kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z\") pod \"d24e47c9-fab9-49dc-9fb3-728e9611a342\" (UID: \"d24e47c9-fab9-49dc-9fb3-728e9611a342\") " Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.250393 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z" (OuterVolumeSpecName: "kube-api-access-w9n8z") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "kube-api-access-w9n8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.253204 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.253311 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph" (OuterVolumeSpecName: "ceph") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.279122 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.285858 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.294212 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory" (OuterVolumeSpecName: "inventory") pod "d24e47c9-fab9-49dc-9fb3-728e9611a342" (UID: "d24e47c9-fab9-49dc-9fb3-728e9611a342"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347623 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347682 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347695 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9n8z\" (UniqueName: \"kubernetes.io/projected/d24e47c9-fab9-49dc-9fb3-728e9611a342-kube-api-access-w9n8z\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347709 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347721 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.347732 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24e47c9-fab9-49dc-9fb3-728e9611a342-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.719924 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" event={"ID":"d24e47c9-fab9-49dc-9fb3-728e9611a342","Type":"ContainerDied","Data":"5086d3104965445ec352f1d87492cd69cbb6f74cbe43272055a7544325951e52"} Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.719963 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5086d3104965445ec352f1d87492cd69cbb6f74cbe43272055a7544325951e52" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.719992 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5hmfm" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.845116 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r"] Oct 06 11:08:41 crc kubenswrapper[4989]: E1006 11:08:41.845551 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24e47c9-fab9-49dc-9fb3-728e9611a342" containerName="neutron-sriov-openstack-openstack-cell1" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.845568 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24e47c9-fab9-49dc-9fb3-728e9611a342" containerName="neutron-sriov-openstack-openstack-cell1" Oct 06 11:08:41 crc kubenswrapper[4989]: E1006 11:08:41.845597 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="extract-content" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.845604 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="extract-content" Oct 06 11:08:41 crc kubenswrapper[4989]: E1006 11:08:41.845626 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="extract-utilities" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.845633 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="extract-utilities" Oct 06 11:08:41 crc kubenswrapper[4989]: E1006 11:08:41.847077 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="registry-server" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.847101 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="registry-server" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.847387 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="267db010-74d9-441b-9b41-ae1825fef07e" containerName="registry-server" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.847434 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24e47c9-fab9-49dc-9fb3-728e9611a342" containerName="neutron-sriov-openstack-openstack-cell1" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.848624 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.850676 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.850845 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.851034 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.852007 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.852642 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.858496 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r"] Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.956877 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xfh\" (UniqueName: \"kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.956925 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.957535 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.957563 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.957604 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:41 crc kubenswrapper[4989]: I1006 11:08:41.957624 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.060067 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.060338 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.060471 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.060916 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.061259 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xfh\" (UniqueName: \"kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.061394 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.063510 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.063720 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.065427 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.065868 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.066546 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.087062 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xfh\" (UniqueName: \"kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh\") pod \"neutron-dhcp-openstack-openstack-cell1-jhx2r\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.178290 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:08:42 crc kubenswrapper[4989]: I1006 11:08:42.807246 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r"] Oct 06 11:08:43 crc kubenswrapper[4989]: I1006 11:08:43.752982 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" event={"ID":"1f292596-66c1-485a-bd87-31d9836fcd48","Type":"ContainerStarted","Data":"6694fd9cf02a6b6af924173d1b2138f95fc9628285e86ba30560ff7896f3ebce"} Oct 06 11:08:43 crc kubenswrapper[4989]: I1006 11:08:43.936754 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:08:43 crc kubenswrapper[4989]: E1006 11:08:43.937375 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:08:44 crc kubenswrapper[4989]: I1006 11:08:44.764250 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" event={"ID":"1f292596-66c1-485a-bd87-31d9836fcd48","Type":"ContainerStarted","Data":"7c2951c9a35de696b8ae6b5dd1379f8d4c114347f1860a0a882b163a1ab1f507"} Oct 06 11:08:44 crc kubenswrapper[4989]: I1006 11:08:44.786018 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" podStartSLOduration=3.642487726 podStartE2EDuration="3.78599981s" podCreationTimestamp="2025-10-06 11:08:41 +0000 UTC" firstStartedPulling="2025-10-06 11:08:42.81974609 +0000 UTC m=+8973.609771670" lastFinishedPulling="2025-10-06 11:08:42.963258164 +0000 UTC m=+8973.753283754" observedRunningTime="2025-10-06 11:08:44.78390611 +0000 UTC m=+8975.573931690" watchObservedRunningTime="2025-10-06 11:08:44.78599981 +0000 UTC m=+8975.576025390" Oct 06 11:08:57 crc kubenswrapper[4989]: I1006 11:08:57.936969 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:08:57 crc kubenswrapper[4989]: E1006 11:08:57.938501 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:09:11 crc kubenswrapper[4989]: I1006 11:09:11.937786 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:09:11 crc kubenswrapper[4989]: E1006 11:09:11.944985 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:09:25 crc kubenswrapper[4989]: I1006 11:09:25.936530 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:09:25 crc kubenswrapper[4989]: E1006 11:09:25.937522 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:09:37 crc kubenswrapper[4989]: I1006 11:09:37.936296 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:09:37 crc kubenswrapper[4989]: E1006 11:09:37.937108 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:09:49 crc kubenswrapper[4989]: I1006 11:09:49.946527 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:09:49 crc kubenswrapper[4989]: E1006 11:09:49.947449 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:10:04 crc kubenswrapper[4989]: I1006 11:10:04.936884 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:10:04 crc kubenswrapper[4989]: E1006 11:10:04.937759 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:10:17 crc kubenswrapper[4989]: I1006 11:10:17.937460 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:10:17 crc kubenswrapper[4989]: E1006 11:10:17.938614 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:10:28 crc kubenswrapper[4989]: I1006 11:10:28.936502 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:10:28 crc kubenswrapper[4989]: E1006 11:10:28.937930 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:10:41 crc kubenswrapper[4989]: I1006 11:10:41.937183 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:10:41 crc kubenswrapper[4989]: E1006 11:10:41.938036 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:10:54 crc kubenswrapper[4989]: I1006 11:10:54.937211 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:10:54 crc kubenswrapper[4989]: E1006 11:10:54.938647 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:07 crc kubenswrapper[4989]: I1006 11:11:07.936100 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:11:07 crc kubenswrapper[4989]: E1006 11:11:07.937060 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:13 crc kubenswrapper[4989]: I1006 11:11:13.514581 4989 generic.go:334] "Generic (PLEG): container finished" podID="1f292596-66c1-485a-bd87-31d9836fcd48" containerID="7c2951c9a35de696b8ae6b5dd1379f8d4c114347f1860a0a882b163a1ab1f507" exitCode=0 Oct 06 11:11:13 crc kubenswrapper[4989]: I1006 11:11:13.514705 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" event={"ID":"1f292596-66c1-485a-bd87-31d9836fcd48","Type":"ContainerDied","Data":"7c2951c9a35de696b8ae6b5dd1379f8d4c114347f1860a0a882b163a1ab1f507"} Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.048899 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.154916 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57xfh\" (UniqueName: \"kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.155268 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.155520 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.156013 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.156159 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.156432 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.161164 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.161273 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph" (OuterVolumeSpecName: "ceph") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.161546 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh" (OuterVolumeSpecName: "kube-api-access-57xfh") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "kube-api-access-57xfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: E1006 11:11:15.187219 4989 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory podName:1f292596-66c1-485a-bd87-31d9836fcd48 nodeName:}" failed. No retries permitted until 2025-10-06 11:11:15.687178574 +0000 UTC m=+9126.477204144 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48") : error deleting /var/lib/kubelet/pods/1f292596-66c1-485a-bd87-31d9836fcd48/volume-subpaths: remove /var/lib/kubelet/pods/1f292596-66c1-485a-bd87-31d9836fcd48/volume-subpaths: no such file or directory Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.187477 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.189397 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.259350 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.259382 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.259393 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.259402 4989 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.259412 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57xfh\" (UniqueName: \"kubernetes.io/projected/1f292596-66c1-485a-bd87-31d9836fcd48-kube-api-access-57xfh\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.536898 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" event={"ID":"1f292596-66c1-485a-bd87-31d9836fcd48","Type":"ContainerDied","Data":"6694fd9cf02a6b6af924173d1b2138f95fc9628285e86ba30560ff7896f3ebce"} Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.537216 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6694fd9cf02a6b6af924173d1b2138f95fc9628285e86ba30560ff7896f3ebce" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.536931 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jhx2r" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.768641 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") pod \"1f292596-66c1-485a-bd87-31d9836fcd48\" (UID: \"1f292596-66c1-485a-bd87-31d9836fcd48\") " Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.779054 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory" (OuterVolumeSpecName: "inventory") pod "1f292596-66c1-485a-bd87-31d9836fcd48" (UID: "1f292596-66c1-485a-bd87-31d9836fcd48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:15 crc kubenswrapper[4989]: I1006 11:11:15.871331 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f292596-66c1-485a-bd87-31d9836fcd48-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:18 crc kubenswrapper[4989]: I1006 11:11:18.936184 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:11:18 crc kubenswrapper[4989]: E1006 11:11:18.937025 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:32 crc kubenswrapper[4989]: I1006 11:11:32.936386 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:11:32 crc kubenswrapper[4989]: E1006 11:11:32.937700 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:40 crc kubenswrapper[4989]: I1006 11:11:40.429765 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:40 crc kubenswrapper[4989]: I1006 11:11:40.430536 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="7d03995f-8dfe-424a-a482-213b5d3b5585" containerName="nova-cell0-conductor-conductor" containerID="cri-o://fef29ab3688c3a740816402adee008be8f7c5dd96401d75cd7449b94c3bee53e" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.453900 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.454317 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="93484eea-55ba-414b-b7e2-a969439587f4" containerName="nova-cell1-conductor-conductor" containerID="cri-o://4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.594093 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.594367 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerName="nova-scheduler-scheduler" containerID="cri-o://34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.606875 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.607158 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-log" containerID="cri-o://49715fe8fe724fc16a887d8425cb2688792ec1ef1b047be7fbff4eb82409acd4" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.607224 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-api" containerID="cri-o://5545728d23140fb4a357835b2146bb2fd1b83a84a4ec6d677315b34e3932d53c" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.620165 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.620405 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" containerID="cri-o://3f2141d4fcbcd8cea4a2c0c4c90b8dbe7553a806e33c6d3f7b6c3a9388fdbcfa" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.620462 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" containerID="cri-o://8a8a4a26f64508dd625d868e89d5a5a400480e11183c70bd21c8cc5895e17d2c" gracePeriod=30 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.818342 4989 generic.go:334] "Generic (PLEG): container finished" podID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerID="3f2141d4fcbcd8cea4a2c0c4c90b8dbe7553a806e33c6d3f7b6c3a9388fdbcfa" exitCode=143 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.818410 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerDied","Data":"3f2141d4fcbcd8cea4a2c0c4c90b8dbe7553a806e33c6d3f7b6c3a9388fdbcfa"} Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.820483 4989 generic.go:334] "Generic (PLEG): container finished" podID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerID="49715fe8fe724fc16a887d8425cb2688792ec1ef1b047be7fbff4eb82409acd4" exitCode=143 Oct 06 11:11:41 crc kubenswrapper[4989]: I1006 11:11:41.820513 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerDied","Data":"49715fe8fe724fc16a887d8425cb2688792ec1ef1b047be7fbff4eb82409acd4"} Oct 06 11:11:42 crc kubenswrapper[4989]: I1006 11:11:42.843492 4989 generic.go:334] "Generic (PLEG): container finished" podID="7d03995f-8dfe-424a-a482-213b5d3b5585" containerID="fef29ab3688c3a740816402adee008be8f7c5dd96401d75cd7449b94c3bee53e" exitCode=0 Oct 06 11:11:42 crc kubenswrapper[4989]: I1006 11:11:42.843540 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d03995f-8dfe-424a-a482-213b5d3b5585","Type":"ContainerDied","Data":"fef29ab3688c3a740816402adee008be8f7c5dd96401d75cd7449b94c3bee53e"} Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.088822 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.116542 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle\") pod \"7d03995f-8dfe-424a-a482-213b5d3b5585\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.116681 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data\") pod \"7d03995f-8dfe-424a-a482-213b5d3b5585\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.116730 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zddlj\" (UniqueName: \"kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj\") pod \"7d03995f-8dfe-424a-a482-213b5d3b5585\" (UID: \"7d03995f-8dfe-424a-a482-213b5d3b5585\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.123296 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj" (OuterVolumeSpecName: "kube-api-access-zddlj") pod "7d03995f-8dfe-424a-a482-213b5d3b5585" (UID: "7d03995f-8dfe-424a-a482-213b5d3b5585"). InnerVolumeSpecName "kube-api-access-zddlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.159135 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data" (OuterVolumeSpecName: "config-data") pod "7d03995f-8dfe-424a-a482-213b5d3b5585" (UID: "7d03995f-8dfe-424a-a482-213b5d3b5585"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.173465 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d03995f-8dfe-424a-a482-213b5d3b5585" (UID: "7d03995f-8dfe-424a-a482-213b5d3b5585"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.218921 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.218953 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d03995f-8dfe-424a-a482-213b5d3b5585-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.218962 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zddlj\" (UniqueName: \"kubernetes.io/projected/7d03995f-8dfe-424a-a482-213b5d3b5585-kube-api-access-zddlj\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.272397 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db is running failed: container process not found" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.272778 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db is running failed: container process not found" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.273210 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db is running failed: container process not found" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.273377 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="93484eea-55ba-414b-b7e2-a969439587f4" containerName="nova-cell1-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.633711 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.828610 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnf7b\" (UniqueName: \"kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b\") pod \"93484eea-55ba-414b-b7e2-a969439587f4\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.828812 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle\") pod \"93484eea-55ba-414b-b7e2-a969439587f4\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.829571 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data\") pod \"93484eea-55ba-414b-b7e2-a969439587f4\" (UID: \"93484eea-55ba-414b-b7e2-a969439587f4\") " Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.833604 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b" (OuterVolumeSpecName: "kube-api-access-lnf7b") pod "93484eea-55ba-414b-b7e2-a969439587f4" (UID: "93484eea-55ba-414b-b7e2-a969439587f4"). InnerVolumeSpecName "kube-api-access-lnf7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.857830 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93484eea-55ba-414b-b7e2-a969439587f4" (UID: "93484eea-55ba-414b-b7e2-a969439587f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.859467 4989 generic.go:334] "Generic (PLEG): container finished" podID="93484eea-55ba-414b-b7e2-a969439587f4" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" exitCode=0 Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.859516 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.859534 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93484eea-55ba-414b-b7e2-a969439587f4","Type":"ContainerDied","Data":"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db"} Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.859573 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"93484eea-55ba-414b-b7e2-a969439587f4","Type":"ContainerDied","Data":"04af45d9118e580b61e613a1eeb7bba118ea0db6a3245cf05bc00fe8f02a51d2"} Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.859604 4989 scope.go:117] "RemoveContainer" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.861143 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data" (OuterVolumeSpecName: "config-data") pod "93484eea-55ba-414b-b7e2-a969439587f4" (UID: "93484eea-55ba-414b-b7e2-a969439587f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.861372 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d03995f-8dfe-424a-a482-213b5d3b5585","Type":"ContainerDied","Data":"58257d0009f1d426760ceaa52c5d7483782b81bca79776432b7b9d547702f028"} Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.861430 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.900824 4989 scope.go:117] "RemoveContainer" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.901227 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db\": container with ID starting with 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db not found: ID does not exist" containerID="4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.901346 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db"} err="failed to get container status \"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db\": rpc error: code = NotFound desc = could not find container \"4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db\": container with ID starting with 4b998b59fd0be1c55c2140bcc030d868e7262f2e47712cc7c1afb5e14a6698db not found: ID does not exist" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.901439 4989 scope.go:117] "RemoveContainer" containerID="fef29ab3688c3a740816402adee008be8f7c5dd96401d75cd7449b94c3bee53e" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.933725 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnf7b\" (UniqueName: \"kubernetes.io/projected/93484eea-55ba-414b-b7e2-a969439587f4-kube-api-access-lnf7b\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.933761 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.933775 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93484eea-55ba-414b-b7e2-a969439587f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.955008 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.955050 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.981343 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.982037 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f292596-66c1-485a-bd87-31d9836fcd48" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982056 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f292596-66c1-485a-bd87-31d9836fcd48" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.982070 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93484eea-55ba-414b-b7e2-a969439587f4" containerName="nova-cell1-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982076 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="93484eea-55ba-414b-b7e2-a969439587f4" containerName="nova-cell1-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: E1006 11:11:43.982087 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d03995f-8dfe-424a-a482-213b5d3b5585" containerName="nova-cell0-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982094 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d03995f-8dfe-424a-a482-213b5d3b5585" containerName="nova-cell0-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982301 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f292596-66c1-485a-bd87-31d9836fcd48" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982312 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d03995f-8dfe-424a-a482-213b5d3b5585" containerName="nova-cell0-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.982336 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="93484eea-55ba-414b-b7e2-a969439587f4" containerName="nova-cell1-conductor-conductor" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.983163 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.985265 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 06 11:11:43 crc kubenswrapper[4989]: I1006 11:11:43.992259 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.036815 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.038027 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.038224 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdh7j\" (UniqueName: \"kubernetes.io/projected/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-kube-api-access-hdh7j\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.140691 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.140774 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.140799 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdh7j\" (UniqueName: \"kubernetes.io/projected/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-kube-api-access-hdh7j\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.145186 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.146129 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.162413 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdh7j\" (UniqueName: \"kubernetes.io/projected/3a757cac-8dd8-42c8-9cdc-4cf3533b9f77-kube-api-access-hdh7j\") pod \"nova-cell0-conductor-0\" (UID: \"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77\") " pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.266740 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.294261 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.316473 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.318062 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.320171 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.326268 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.344168 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9xvq\" (UniqueName: \"kubernetes.io/projected/301be596-d22c-4d61-9436-7f93b25bbd8a-kube-api-access-b9xvq\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.344357 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.344407 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.369897 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.447011 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.447112 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.447246 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9xvq\" (UniqueName: \"kubernetes.io/projected/301be596-d22c-4d61-9436-7f93b25bbd8a-kube-api-access-b9xvq\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.454349 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.456220 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/301be596-d22c-4d61-9436-7f93b25bbd8a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.484276 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9xvq\" (UniqueName: \"kubernetes.io/projected/301be596-d22c-4d61-9436-7f93b25bbd8a-kube-api-access-b9xvq\") pod \"nova-cell1-conductor-0\" (UID: \"301be596-d22c-4d61-9436-7f93b25bbd8a\") " pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.640234 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.820801 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:54412->10.217.1.81:8775: read: connection reset by peer" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.820802 4989 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.81:8775/\": read tcp 10.217.0.2:54410->10.217.1.81:8775: read: connection reset by peer" Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.934052 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 06 11:11:44 crc kubenswrapper[4989]: I1006 11:11:44.936443 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:11:44 crc kubenswrapper[4989]: E1006 11:11:44.937027 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.934095 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77","Type":"ContainerStarted","Data":"3048f35067d70b8ab269f6db501e66dca9603161b480097f9e9b9cc3a54f5e1d"} Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.934721 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3a757cac-8dd8-42c8-9cdc-4cf3533b9f77","Type":"ContainerStarted","Data":"f331bbe425893a0c8679db27de695a41543a0b9d37f28044db8cbc9567394cc6"} Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.942414 4989 generic.go:334] "Generic (PLEG): container finished" podID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerID="5545728d23140fb4a357835b2146bb2fd1b83a84a4ec6d677315b34e3932d53c" exitCode=0 Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.946007 4989 generic.go:334] "Generic (PLEG): container finished" podID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerID="8a8a4a26f64508dd625d868e89d5a5a400480e11183c70bd21c8cc5895e17d2c" exitCode=0 Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.946956 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d03995f-8dfe-424a-a482-213b5d3b5585" path="/var/lib/kubelet/pods/7d03995f-8dfe-424a-a482-213b5d3b5585/volumes" Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.947405 4989 generic.go:334] "Generic (PLEG): container finished" podID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerID="34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" exitCode=0 Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.948249 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93484eea-55ba-414b-b7e2-a969439587f4" path="/var/lib/kubelet/pods/93484eea-55ba-414b-b7e2-a969439587f4/volumes" Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.949496 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerDied","Data":"5545728d23140fb4a357835b2146bb2fd1b83a84a4ec6d677315b34e3932d53c"} Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.949530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerDied","Data":"8a8a4a26f64508dd625d868e89d5a5a400480e11183c70bd21c8cc5895e17d2c"} Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.949545 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15","Type":"ContainerDied","Data":"34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823"} Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.949564 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:45 crc kubenswrapper[4989]: I1006 11:11:45.965168 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.965145642 podStartE2EDuration="2.965145642s" podCreationTimestamp="2025-10-06 11:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:11:45.96122867 +0000 UTC m=+9156.751254260" watchObservedRunningTime="2025-10-06 11:11:45.965145642 +0000 UTC m=+9156.755171222" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.116349 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 06 11:11:46 crc kubenswrapper[4989]: W1006 11:11:46.120981 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod301be596_d22c_4d61_9436_7f93b25bbd8a.slice/crio-d2f949fd37a424926273531a78357a5f71306fe6e0442712ebcb2ff081ca5002 WatchSource:0}: Error finding container d2f949fd37a424926273531a78357a5f71306fe6e0442712ebcb2ff081ca5002: Status 404 returned error can't find the container with id d2f949fd37a424926273531a78357a5f71306fe6e0442712ebcb2ff081ca5002 Oct 06 11:11:46 crc kubenswrapper[4989]: E1006 11:11:46.142960 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823 is running failed: container process not found" containerID="34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 11:11:46 crc kubenswrapper[4989]: E1006 11:11:46.145069 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823 is running failed: container process not found" containerID="34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 11:11:46 crc kubenswrapper[4989]: E1006 11:11:46.146295 4989 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823 is running failed: container process not found" containerID="34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 06 11:11:46 crc kubenswrapper[4989]: E1006 11:11:46.146333 4989 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerName="nova-scheduler-scheduler" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.240699 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.248132 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.274628 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411139 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpfmb\" (UniqueName: \"kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb\") pod \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411482 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs\") pod \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411522 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle\") pod \"5c24cb4b-6793-4479-86e5-fd42c3a91717\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dhxc\" (UniqueName: \"kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc\") pod \"5c24cb4b-6793-4479-86e5-fd42c3a91717\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411612 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data\") pod \"5c24cb4b-6793-4479-86e5-fd42c3a91717\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411798 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqbsz\" (UniqueName: \"kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz\") pod \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411849 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle\") pod \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.411935 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs\") pod \"5c24cb4b-6793-4479-86e5-fd42c3a91717\" (UID: \"5c24cb4b-6793-4479-86e5-fd42c3a91717\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.412043 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data\") pod \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\" (UID: \"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.412088 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data\") pod \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.412135 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle\") pod \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\" (UID: \"81438bf6-a32a-443e-a94a-b6882c9b8ca3\") " Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.417004 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs" (OuterVolumeSpecName: "logs") pod "5c24cb4b-6793-4479-86e5-fd42c3a91717" (UID: "5c24cb4b-6793-4479-86e5-fd42c3a91717"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.418028 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs" (OuterVolumeSpecName: "logs") pod "81438bf6-a32a-443e-a94a-b6882c9b8ca3" (UID: "81438bf6-a32a-443e-a94a-b6882c9b8ca3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.425330 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc" (OuterVolumeSpecName: "kube-api-access-6dhxc") pod "5c24cb4b-6793-4479-86e5-fd42c3a91717" (UID: "5c24cb4b-6793-4479-86e5-fd42c3a91717"). InnerVolumeSpecName "kube-api-access-6dhxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.425408 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb" (OuterVolumeSpecName: "kube-api-access-zpfmb") pod "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" (UID: "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15"). InnerVolumeSpecName "kube-api-access-zpfmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.426814 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz" (OuterVolumeSpecName: "kube-api-access-gqbsz") pod "81438bf6-a32a-443e-a94a-b6882c9b8ca3" (UID: "81438bf6-a32a-443e-a94a-b6882c9b8ca3"). InnerVolumeSpecName "kube-api-access-gqbsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.468496 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" (UID: "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.478629 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81438bf6-a32a-443e-a94a-b6882c9b8ca3" (UID: "81438bf6-a32a-443e-a94a-b6882c9b8ca3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517716 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517756 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpfmb\" (UniqueName: \"kubernetes.io/projected/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-kube-api-access-zpfmb\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517769 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81438bf6-a32a-443e-a94a-b6882c9b8ca3-logs\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517779 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dhxc\" (UniqueName: \"kubernetes.io/projected/5c24cb4b-6793-4479-86e5-fd42c3a91717-kube-api-access-6dhxc\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517788 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqbsz\" (UniqueName: \"kubernetes.io/projected/81438bf6-a32a-443e-a94a-b6882c9b8ca3-kube-api-access-gqbsz\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517797 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.517805 4989 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c24cb4b-6793-4479-86e5-fd42c3a91717-logs\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.519081 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data" (OuterVolumeSpecName: "config-data") pod "5c24cb4b-6793-4479-86e5-fd42c3a91717" (UID: "5c24cb4b-6793-4479-86e5-fd42c3a91717"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.520783 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data" (OuterVolumeSpecName: "config-data") pod "81438bf6-a32a-443e-a94a-b6882c9b8ca3" (UID: "81438bf6-a32a-443e-a94a-b6882c9b8ca3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.531507 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c24cb4b-6793-4479-86e5-fd42c3a91717" (UID: "5c24cb4b-6793-4479-86e5-fd42c3a91717"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.548021 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data" (OuterVolumeSpecName: "config-data") pod "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" (UID: "4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.621283 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.621317 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81438bf6-a32a-443e-a94a-b6882c9b8ca3-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.621326 4989 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.621337 4989 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c24cb4b-6793-4479-86e5-fd42c3a91717-config-data\") on node \"crc\" DevicePath \"\"" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.972046 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5c24cb4b-6793-4479-86e5-fd42c3a91717","Type":"ContainerDied","Data":"f97986c8414427053b9e48c4b982210157585ec9b85d28f218d6ae22d6e9c026"} Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.972123 4989 scope.go:117] "RemoveContainer" containerID="8a8a4a26f64508dd625d868e89d5a5a400480e11183c70bd21c8cc5895e17d2c" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.972345 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.978870 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.979288 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15","Type":"ContainerDied","Data":"dcb2dd865fe7fe2bd76e3993478ac09c3728da938247ead7cc59e8480e0d2f24"} Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.983004 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"301be596-d22c-4d61-9436-7f93b25bbd8a","Type":"ContainerStarted","Data":"4a9bb07d150cb558ba8dc4b9f6588860a0c5e66c1d9916635aa3cfd3283902b2"} Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.983056 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"301be596-d22c-4d61-9436-7f93b25bbd8a","Type":"ContainerStarted","Data":"d2f949fd37a424926273531a78357a5f71306fe6e0442712ebcb2ff081ca5002"} Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.983988 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.986926 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 11:11:46 crc kubenswrapper[4989]: I1006 11:11:46.986979 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81438bf6-a32a-443e-a94a-b6882c9b8ca3","Type":"ContainerDied","Data":"7559f357d36fb81e79f90d49151a4737fe312b60003d91fcff3604f39262742c"} Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.011999 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.011974853 podStartE2EDuration="3.011974853s" podCreationTimestamp="2025-10-06 11:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:11:47.004810036 +0000 UTC m=+9157.794835616" watchObservedRunningTime="2025-10-06 11:11:47.011974853 +0000 UTC m=+9157.802000433" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.581892 4989 scope.go:117] "RemoveContainer" containerID="3f2141d4fcbcd8cea4a2c0c4c90b8dbe7553a806e33c6d3f7b6c3a9388fdbcfa" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.584264 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.599392 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.625280 4989 scope.go:117] "RemoveContainer" containerID="34fdde3765c7208700c5b1eeb2bfcf09cc068351fc0ff2617003a892d967f823" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.631373 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.644719 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.658317 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: E1006 11:11:47.659323 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-log" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659344 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-log" Oct 06 11:11:47 crc kubenswrapper[4989]: E1006 11:11:47.659369 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659375 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" Oct 06 11:11:47 crc kubenswrapper[4989]: E1006 11:11:47.659382 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659389 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" Oct 06 11:11:47 crc kubenswrapper[4989]: E1006 11:11:47.659403 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerName="nova-scheduler-scheduler" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659410 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerName="nova-scheduler-scheduler" Oct 06 11:11:47 crc kubenswrapper[4989]: E1006 11:11:47.659433 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-api" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659441 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-api" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659711 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-log" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659723 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" containerName="nova-metadata-metadata" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659741 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-log" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659758 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" containerName="nova-scheduler-scheduler" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.659769 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" containerName="nova-api-api" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.661026 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.664052 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.683707 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.696939 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.705449 4989 scope.go:117] "RemoveContainer" containerID="5545728d23140fb4a357835b2146bb2fd1b83a84a4ec6d677315b34e3932d53c" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.724856 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.727224 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.729583 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.729898 4989 scope.go:117] "RemoveContainer" containerID="49715fe8fe724fc16a887d8425cb2688792ec1ef1b047be7fbff4eb82409acd4" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.730051 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.731981 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.733782 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.742389 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.746695 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/198f79e4-9643-4782-b5b8-269d0b51d986-kube-api-access-gbxsx\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747114 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198f79e4-9643-4782-b5b8-269d0b51d986-logs\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747230 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7x55\" (UniqueName: \"kubernetes.io/projected/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-kube-api-access-l7x55\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747262 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-config-data\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-config-data\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747340 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747387 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.747411 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-logs\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.749715 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.759039 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849808 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-config-data\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849858 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849898 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849931 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849948 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsrpt\" (UniqueName: \"kubernetes.io/projected/0cd872d1-fade-451e-82bc-14993f72da4a-kube-api-access-vsrpt\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.849976 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-logs\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850160 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/198f79e4-9643-4782-b5b8-269d0b51d986-kube-api-access-gbxsx\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850289 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198f79e4-9643-4782-b5b8-269d0b51d986-logs\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850500 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-logs\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850545 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7x55\" (UniqueName: \"kubernetes.io/projected/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-kube-api-access-l7x55\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850586 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-config-data\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850664 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-config-data\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.850859 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/198f79e4-9643-4782-b5b8-269d0b51d986-logs\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.855783 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.856451 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-config-data\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.856467 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-config-data\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.858302 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198f79e4-9643-4782-b5b8-269d0b51d986-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.868256 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7x55\" (UniqueName: \"kubernetes.io/projected/430c6cf8-e65b-4c3e-aeec-6a22e2f472b5-kube-api-access-l7x55\") pod \"nova-api-0\" (UID: \"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5\") " pod="openstack/nova-api-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.873908 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxsx\" (UniqueName: \"kubernetes.io/projected/198f79e4-9643-4782-b5b8-269d0b51d986-kube-api-access-gbxsx\") pod \"nova-metadata-0\" (UID: \"198f79e4-9643-4782-b5b8-269d0b51d986\") " pod="openstack/nova-metadata-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.949059 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15" path="/var/lib/kubelet/pods/4f8ba4f7-dc15-4ba2-af9a-8f9faba12a15/volumes" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.952056 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-config-data\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.952136 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.952176 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsrpt\" (UniqueName: \"kubernetes.io/projected/0cd872d1-fade-451e-82bc-14993f72da4a-kube-api-access-vsrpt\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.952995 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c24cb4b-6793-4479-86e5-fd42c3a91717" path="/var/lib/kubelet/pods/5c24cb4b-6793-4479-86e5-fd42c3a91717/volumes" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.954641 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81438bf6-a32a-443e-a94a-b6882c9b8ca3" path="/var/lib/kubelet/pods/81438bf6-a32a-443e-a94a-b6882c9b8ca3/volumes" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.955367 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.955724 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cd872d1-fade-451e-82bc-14993f72da4a-config-data\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.983481 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsrpt\" (UniqueName: \"kubernetes.io/projected/0cd872d1-fade-451e-82bc-14993f72da4a-kube-api-access-vsrpt\") pod \"nova-scheduler-0\" (UID: \"0cd872d1-fade-451e-82bc-14993f72da4a\") " pod="openstack/nova-scheduler-0" Oct 06 11:11:47 crc kubenswrapper[4989]: I1006 11:11:47.984356 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 06 11:11:48 crc kubenswrapper[4989]: I1006 11:11:48.053034 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 06 11:11:48 crc kubenswrapper[4989]: I1006 11:11:48.063141 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 06 11:11:48 crc kubenswrapper[4989]: I1006 11:11:48.545075 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 06 11:11:48 crc kubenswrapper[4989]: I1006 11:11:48.717175 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 06 11:11:48 crc kubenswrapper[4989]: W1006 11:11:48.722518 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cd872d1_fade_451e_82bc_14993f72da4a.slice/crio-dce3a638d6cfef065b9953183166de707fe4db193bccf5fd0fb039a7f8309fb6 WatchSource:0}: Error finding container dce3a638d6cfef065b9953183166de707fe4db193bccf5fd0fb039a7f8309fb6: Status 404 returned error can't find the container with id dce3a638d6cfef065b9953183166de707fe4db193bccf5fd0fb039a7f8309fb6 Oct 06 11:11:48 crc kubenswrapper[4989]: I1006 11:11:48.734344 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.022168 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0cd872d1-fade-451e-82bc-14993f72da4a","Type":"ContainerStarted","Data":"bec471ed4835946f8de8cb4e9420fed813aa192962c2c28ced3a5020c205ad77"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.022458 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0cd872d1-fade-451e-82bc-14993f72da4a","Type":"ContainerStarted","Data":"dce3a638d6cfef065b9953183166de707fe4db193bccf5fd0fb039a7f8309fb6"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.032367 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198f79e4-9643-4782-b5b8-269d0b51d986","Type":"ContainerStarted","Data":"347b636d619b393cd4b134305b1e2a746e93cc8591e36697671c071ba3529e26"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.032411 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198f79e4-9643-4782-b5b8-269d0b51d986","Type":"ContainerStarted","Data":"3c8078d1c0c6f3573aa19a7cf314a3863e99446fcdc12b9b8e5d1dbe036363ab"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.032422 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"198f79e4-9643-4782-b5b8-269d0b51d986","Type":"ContainerStarted","Data":"f9e9b32d7d0580fe9f120d151f17f441ca9abce9f007baf84df7c892ef41f44c"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.034790 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5","Type":"ContainerStarted","Data":"a7a281fce446ddca78dbcdee166457abda3d54d06ad1d3b9e57949203eda1a81"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.034846 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5","Type":"ContainerStarted","Data":"185737a504865a37ff031ff8d337d76b0571ac1f4e9edad0da8bd4e61f06c223"} Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.047760 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.047742178 podStartE2EDuration="2.047742178s" podCreationTimestamp="2025-10-06 11:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:11:49.043014611 +0000 UTC m=+9159.833040211" watchObservedRunningTime="2025-10-06 11:11:49.047742178 +0000 UTC m=+9159.837767758" Oct 06 11:11:49 crc kubenswrapper[4989]: I1006 11:11:49.071704 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.071682637 podStartE2EDuration="2.071682637s" podCreationTimestamp="2025-10-06 11:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:11:49.061772032 +0000 UTC m=+9159.851797622" watchObservedRunningTime="2025-10-06 11:11:49.071682637 +0000 UTC m=+9159.861708227" Oct 06 11:11:50 crc kubenswrapper[4989]: I1006 11:11:50.046702 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"430c6cf8-e65b-4c3e-aeec-6a22e2f472b5","Type":"ContainerStarted","Data":"90158dd9b777cd9081add93755678073f7818f43e00d2978cdb90196bcea2fff"} Oct 06 11:11:50 crc kubenswrapper[4989]: I1006 11:11:50.084524 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.084503279 podStartE2EDuration="3.084503279s" podCreationTimestamp="2025-10-06 11:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:11:50.073006758 +0000 UTC m=+9160.863032338" watchObservedRunningTime="2025-10-06 11:11:50.084503279 +0000 UTC m=+9160.874528859" Oct 06 11:11:53 crc kubenswrapper[4989]: I1006 11:11:53.002956 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 11:11:53 crc kubenswrapper[4989]: I1006 11:11:53.003675 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 06 11:11:53 crc kubenswrapper[4989]: I1006 11:11:53.053514 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 06 11:11:54 crc kubenswrapper[4989]: I1006 11:11:54.426621 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 06 11:11:54 crc kubenswrapper[4989]: I1006 11:11:54.710051 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 06 11:11:55 crc kubenswrapper[4989]: I1006 11:11:55.936962 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:11:55 crc kubenswrapper[4989]: E1006 11:11:55.937362 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:11:57 crc kubenswrapper[4989]: I1006 11:11:57.984794 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 11:11:57 crc kubenswrapper[4989]: I1006 11:11:57.985693 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 06 11:11:58 crc kubenswrapper[4989]: I1006 11:11:58.054257 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 06 11:11:58 crc kubenswrapper[4989]: I1006 11:11:58.063451 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 11:11:58 crc kubenswrapper[4989]: I1006 11:11:58.063497 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 06 11:11:58 crc kubenswrapper[4989]: I1006 11:11:58.090415 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 06 11:11:58 crc kubenswrapper[4989]: I1006 11:11:58.196178 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 06 11:11:59 crc kubenswrapper[4989]: I1006 11:11:59.067863 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="198f79e4-9643-4782-b5b8-269d0b51d986" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 11:11:59 crc kubenswrapper[4989]: I1006 11:11:59.067960 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="198f79e4-9643-4782-b5b8-269d0b51d986" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.188:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 11:11:59 crc kubenswrapper[4989]: I1006 11:11:59.149877 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="430c6cf8-e65b-4c3e-aeec-6a22e2f472b5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 11:11:59 crc kubenswrapper[4989]: I1006 11:11:59.150145 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="430c6cf8-e65b-4c3e-aeec-6a22e2f472b5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.190:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 06 11:12:06 crc kubenswrapper[4989]: I1006 11:12:06.936536 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:12:06 crc kubenswrapper[4989]: E1006 11:12:06.938247 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:12:07 crc kubenswrapper[4989]: I1006 11:12:07.987374 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 11:12:07 crc kubenswrapper[4989]: I1006 11:12:07.987938 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 06 11:12:07 crc kubenswrapper[4989]: I1006 11:12:07.991896 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 11:12:07 crc kubenswrapper[4989]: I1006 11:12:07.994036 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.084342 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.085029 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.085069 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.089735 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.287105 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 06 11:12:08 crc kubenswrapper[4989]: I1006 11:12:08.290904 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.662377 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw"] Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.666959 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.672931 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.672979 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.673082 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.673124 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.673238 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.673584 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.673709 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.679483 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw"] Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771254 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771302 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwspj\" (UniqueName: \"kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771336 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771370 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771390 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771473 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771494 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771513 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771542 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771572 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.771604 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.873858 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.873931 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwspj\" (UniqueName: \"kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874004 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874083 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874142 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874378 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874483 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874543 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874589 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874631 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.874699 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.876075 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.876277 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.881781 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.881820 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.882070 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.882135 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.882332 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.883302 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.885098 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.889333 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:09 crc kubenswrapper[4989]: I1006 11:12:09.893280 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwspj\" (UniqueName: \"kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:10 crc kubenswrapper[4989]: I1006 11:12:10.023258 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-tgbxk" Oct 06 11:12:10 crc kubenswrapper[4989]: I1006 11:12:10.031770 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:12:10 crc kubenswrapper[4989]: I1006 11:12:10.590667 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw"] Oct 06 11:12:10 crc kubenswrapper[4989]: I1006 11:12:10.790388 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 06 11:12:11 crc kubenswrapper[4989]: I1006 11:12:11.321438 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" event={"ID":"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4","Type":"ContainerStarted","Data":"1962ff8fd838da3d54fd9f7c35699f137f3e430cdc90b629ff47f877e5d6643c"} Oct 06 11:12:12 crc kubenswrapper[4989]: I1006 11:12:12.337360 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" event={"ID":"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4","Type":"ContainerStarted","Data":"5bc08c291a71b0cc6cecbf6b898d85f37d2db2fabfedbcf8f2628ddc34448dd5"} Oct 06 11:12:12 crc kubenswrapper[4989]: I1006 11:12:12.375208 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" podStartSLOduration=3.184711572 podStartE2EDuration="3.37518791s" podCreationTimestamp="2025-10-06 11:12:09 +0000 UTC" firstStartedPulling="2025-10-06 11:12:10.596299548 +0000 UTC m=+9181.386325138" lastFinishedPulling="2025-10-06 11:12:10.786775896 +0000 UTC m=+9181.576801476" observedRunningTime="2025-10-06 11:12:12.363809002 +0000 UTC m=+9183.153834592" watchObservedRunningTime="2025-10-06 11:12:12.37518791 +0000 UTC m=+9183.165213490" Oct 06 11:12:17 crc kubenswrapper[4989]: I1006 11:12:17.937881 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:12:17 crc kubenswrapper[4989]: E1006 11:12:17.939168 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:12:28 crc kubenswrapper[4989]: I1006 11:12:28.936140 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:12:28 crc kubenswrapper[4989]: E1006 11:12:28.936969 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:12:41 crc kubenswrapper[4989]: I1006 11:12:41.935838 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:12:42 crc kubenswrapper[4989]: I1006 11:12:42.730859 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c"} Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.708945 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.711848 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.737303 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.877560 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.877626 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.877648 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cthnx\" (UniqueName: \"kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.979433 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.979510 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cthnx\" (UniqueName: \"kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.979531 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.980164 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:12:59 crc kubenswrapper[4989]: I1006 11:12:59.980161 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.004482 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cthnx\" (UniqueName: \"kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx\") pod \"redhat-marketplace-7t8wr\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.043307 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.584069 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.946624 4989 generic.go:334] "Generic (PLEG): container finished" podID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerID="2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057" exitCode=0 Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.946701 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerDied","Data":"2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057"} Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.947012 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerStarted","Data":"5d664394b6a9bd093351e950b43c741d25b3b7389350c13446eb320e0d405f34"} Oct 06 11:13:00 crc kubenswrapper[4989]: I1006 11:13:00.948431 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:13:01 crc kubenswrapper[4989]: I1006 11:13:01.960245 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerStarted","Data":"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e"} Oct 06 11:13:02 crc kubenswrapper[4989]: I1006 11:13:02.972488 4989 generic.go:334] "Generic (PLEG): container finished" podID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerID="34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e" exitCode=0 Oct 06 11:13:02 crc kubenswrapper[4989]: I1006 11:13:02.972548 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerDied","Data":"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e"} Oct 06 11:13:03 crc kubenswrapper[4989]: I1006 11:13:03.986776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerStarted","Data":"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07"} Oct 06 11:13:04 crc kubenswrapper[4989]: I1006 11:13:04.004586 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7t8wr" podStartSLOduration=2.469579218 podStartE2EDuration="5.004565015s" podCreationTimestamp="2025-10-06 11:12:59 +0000 UTC" firstStartedPulling="2025-10-06 11:13:00.948189466 +0000 UTC m=+9231.738215056" lastFinishedPulling="2025-10-06 11:13:03.483175263 +0000 UTC m=+9234.273200853" observedRunningTime="2025-10-06 11:13:04.002322621 +0000 UTC m=+9234.792348201" watchObservedRunningTime="2025-10-06 11:13:04.004565015 +0000 UTC m=+9234.794590595" Oct 06 11:13:10 crc kubenswrapper[4989]: I1006 11:13:10.044412 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:10 crc kubenswrapper[4989]: I1006 11:13:10.046567 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:10 crc kubenswrapper[4989]: I1006 11:13:10.120320 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:11 crc kubenswrapper[4989]: I1006 11:13:11.149328 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:11 crc kubenswrapper[4989]: I1006 11:13:11.210432 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.105333 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7t8wr" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="registry-server" containerID="cri-o://19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07" gracePeriod=2 Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.598691 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.727444 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content\") pod \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.727513 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cthnx\" (UniqueName: \"kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx\") pod \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.727899 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities\") pod \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\" (UID: \"85d473ab-f63e-4fc6-9f9e-d48a4eec962b\") " Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.728822 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities" (OuterVolumeSpecName: "utilities") pod "85d473ab-f63e-4fc6-9f9e-d48a4eec962b" (UID: "85d473ab-f63e-4fc6-9f9e-d48a4eec962b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.734372 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx" (OuterVolumeSpecName: "kube-api-access-cthnx") pod "85d473ab-f63e-4fc6-9f9e-d48a4eec962b" (UID: "85d473ab-f63e-4fc6-9f9e-d48a4eec962b"). InnerVolumeSpecName "kube-api-access-cthnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.740145 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85d473ab-f63e-4fc6-9f9e-d48a4eec962b" (UID: "85d473ab-f63e-4fc6-9f9e-d48a4eec962b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.831938 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.832026 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:13:13 crc kubenswrapper[4989]: I1006 11:13:13.832049 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cthnx\" (UniqueName: \"kubernetes.io/projected/85d473ab-f63e-4fc6-9f9e-d48a4eec962b-kube-api-access-cthnx\") on node \"crc\" DevicePath \"\"" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.122131 4989 generic.go:334] "Generic (PLEG): container finished" podID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerID="19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07" exitCode=0 Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.122200 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerDied","Data":"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07"} Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.122221 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7t8wr" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.122240 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7t8wr" event={"ID":"85d473ab-f63e-4fc6-9f9e-d48a4eec962b","Type":"ContainerDied","Data":"5d664394b6a9bd093351e950b43c741d25b3b7389350c13446eb320e0d405f34"} Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.122269 4989 scope.go:117] "RemoveContainer" containerID="19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.179753 4989 scope.go:117] "RemoveContainer" containerID="34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.181998 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.195222 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7t8wr"] Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.219108 4989 scope.go:117] "RemoveContainer" containerID="2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.282342 4989 scope.go:117] "RemoveContainer" containerID="19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07" Oct 06 11:13:14 crc kubenswrapper[4989]: E1006 11:13:14.282862 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07\": container with ID starting with 19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07 not found: ID does not exist" containerID="19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.282978 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07"} err="failed to get container status \"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07\": rpc error: code = NotFound desc = could not find container \"19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07\": container with ID starting with 19fd9418fc6eb97074aafb9b93682355f5fc5c4cc091fcaf1c08543522926a07 not found: ID does not exist" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.283106 4989 scope.go:117] "RemoveContainer" containerID="34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e" Oct 06 11:13:14 crc kubenswrapper[4989]: E1006 11:13:14.283579 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e\": container with ID starting with 34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e not found: ID does not exist" containerID="34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.283609 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e"} err="failed to get container status \"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e\": rpc error: code = NotFound desc = could not find container \"34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e\": container with ID starting with 34693ba857c07a31821dbfabf13d9b1f1e07fb0e2f8bd01949ac25ed9a58c05e not found: ID does not exist" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.283630 4989 scope.go:117] "RemoveContainer" containerID="2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057" Oct 06 11:13:14 crc kubenswrapper[4989]: E1006 11:13:14.284099 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057\": container with ID starting with 2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057 not found: ID does not exist" containerID="2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057" Oct 06 11:13:14 crc kubenswrapper[4989]: I1006 11:13:14.284242 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057"} err="failed to get container status \"2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057\": rpc error: code = NotFound desc = could not find container \"2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057\": container with ID starting with 2d928e2267d57e562238550d3d74e9c83e681ff63cb15be76a5ea9851d421057 not found: ID does not exist" Oct 06 11:13:15 crc kubenswrapper[4989]: I1006 11:13:15.949836 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" path="/var/lib/kubelet/pods/85d473ab-f63e-4fc6-9f9e-d48a4eec962b/volumes" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.280109 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:13:48 crc kubenswrapper[4989]: E1006 11:13:48.281769 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="extract-utilities" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.281794 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="extract-utilities" Oct 06 11:13:48 crc kubenswrapper[4989]: E1006 11:13:48.281859 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="registry-server" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.281872 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="registry-server" Oct 06 11:13:48 crc kubenswrapper[4989]: E1006 11:13:48.281894 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="extract-content" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.281904 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="extract-content" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.282315 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d473ab-f63e-4fc6-9f9e-d48a4eec962b" containerName="registry-server" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.286980 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.300259 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.417979 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.418448 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.418504 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxj8x\" (UniqueName: \"kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.519987 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.520071 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.520127 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxj8x\" (UniqueName: \"kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.521072 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.521257 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.545261 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxj8x\" (UniqueName: \"kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x\") pod \"community-operators-ccjfp\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:48 crc kubenswrapper[4989]: I1006 11:13:48.642594 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:49 crc kubenswrapper[4989]: I1006 11:13:49.221889 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:13:49 crc kubenswrapper[4989]: W1006 11:13:49.223102 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod244a8549_d6d2_4d9b_b1b8_0f089c2436c0.slice/crio-5a8e273c66f120b4e3e63a4b615391e518c90422d0bba146baa4f347a468e2ae WatchSource:0}: Error finding container 5a8e273c66f120b4e3e63a4b615391e518c90422d0bba146baa4f347a468e2ae: Status 404 returned error can't find the container with id 5a8e273c66f120b4e3e63a4b615391e518c90422d0bba146baa4f347a468e2ae Oct 06 11:13:49 crc kubenswrapper[4989]: I1006 11:13:49.597831 4989 generic.go:334] "Generic (PLEG): container finished" podID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerID="65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e" exitCode=0 Oct 06 11:13:49 crc kubenswrapper[4989]: I1006 11:13:49.597925 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerDied","Data":"65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e"} Oct 06 11:13:49 crc kubenswrapper[4989]: I1006 11:13:49.600839 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerStarted","Data":"5a8e273c66f120b4e3e63a4b615391e518c90422d0bba146baa4f347a468e2ae"} Oct 06 11:13:50 crc kubenswrapper[4989]: I1006 11:13:50.627186 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerStarted","Data":"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080"} Oct 06 11:13:51 crc kubenswrapper[4989]: I1006 11:13:51.641564 4989 generic.go:334] "Generic (PLEG): container finished" podID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerID="bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080" exitCode=0 Oct 06 11:13:51 crc kubenswrapper[4989]: I1006 11:13:51.641628 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerDied","Data":"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080"} Oct 06 11:13:53 crc kubenswrapper[4989]: I1006 11:13:53.664841 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerStarted","Data":"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9"} Oct 06 11:13:53 crc kubenswrapper[4989]: I1006 11:13:53.690691 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ccjfp" podStartSLOduration=3.062036305 podStartE2EDuration="5.6906375s" podCreationTimestamp="2025-10-06 11:13:48 +0000 UTC" firstStartedPulling="2025-10-06 11:13:49.600843745 +0000 UTC m=+9280.390869355" lastFinishedPulling="2025-10-06 11:13:52.22944496 +0000 UTC m=+9283.019470550" observedRunningTime="2025-10-06 11:13:53.683392861 +0000 UTC m=+9284.473418461" watchObservedRunningTime="2025-10-06 11:13:53.6906375 +0000 UTC m=+9284.480663110" Oct 06 11:13:58 crc kubenswrapper[4989]: I1006 11:13:58.643585 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:58 crc kubenswrapper[4989]: I1006 11:13:58.644460 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:58 crc kubenswrapper[4989]: I1006 11:13:58.732342 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:58 crc kubenswrapper[4989]: I1006 11:13:58.796920 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:13:58 crc kubenswrapper[4989]: I1006 11:13:58.991749 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:14:00 crc kubenswrapper[4989]: I1006 11:14:00.762580 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ccjfp" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="registry-server" containerID="cri-o://971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9" gracePeriod=2 Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.334052 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.459052 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities\") pod \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.459314 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content\") pod \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.459428 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxj8x\" (UniqueName: \"kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x\") pod \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\" (UID: \"244a8549-d6d2-4d9b-b1b8-0f089c2436c0\") " Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.460107 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities" (OuterVolumeSpecName: "utilities") pod "244a8549-d6d2-4d9b-b1b8-0f089c2436c0" (UID: "244a8549-d6d2-4d9b-b1b8-0f089c2436c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.468887 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x" (OuterVolumeSpecName: "kube-api-access-nxj8x") pod "244a8549-d6d2-4d9b-b1b8-0f089c2436c0" (UID: "244a8549-d6d2-4d9b-b1b8-0f089c2436c0"). InnerVolumeSpecName "kube-api-access-nxj8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.522558 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "244a8549-d6d2-4d9b-b1b8-0f089c2436c0" (UID: "244a8549-d6d2-4d9b-b1b8-0f089c2436c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.562287 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.562312 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.562323 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxj8x\" (UniqueName: \"kubernetes.io/projected/244a8549-d6d2-4d9b-b1b8-0f089c2436c0-kube-api-access-nxj8x\") on node \"crc\" DevicePath \"\"" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.777203 4989 generic.go:334] "Generic (PLEG): container finished" podID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerID="971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9" exitCode=0 Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.777284 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerDied","Data":"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9"} Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.777496 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ccjfp" event={"ID":"244a8549-d6d2-4d9b-b1b8-0f089c2436c0","Type":"ContainerDied","Data":"5a8e273c66f120b4e3e63a4b615391e518c90422d0bba146baa4f347a468e2ae"} Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.777520 4989 scope.go:117] "RemoveContainer" containerID="971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.777368 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ccjfp" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.823254 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.834684 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ccjfp"] Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.844381 4989 scope.go:117] "RemoveContainer" containerID="bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.888346 4989 scope.go:117] "RemoveContainer" containerID="65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.933068 4989 scope.go:117] "RemoveContainer" containerID="971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9" Oct 06 11:14:01 crc kubenswrapper[4989]: E1006 11:14:01.933482 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9\": container with ID starting with 971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9 not found: ID does not exist" containerID="971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.933519 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9"} err="failed to get container status \"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9\": rpc error: code = NotFound desc = could not find container \"971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9\": container with ID starting with 971ea98fb1b76e03c94b6a86cb46b224d26e007b57ca1b1efa3877fa200120e9 not found: ID does not exist" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.933541 4989 scope.go:117] "RemoveContainer" containerID="bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080" Oct 06 11:14:01 crc kubenswrapper[4989]: E1006 11:14:01.934791 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080\": container with ID starting with bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080 not found: ID does not exist" containerID="bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.934813 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080"} err="failed to get container status \"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080\": rpc error: code = NotFound desc = could not find container \"bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080\": container with ID starting with bf75594a90641fc82b3b8f8893d3ca6e20c9cd2411bd4d0c18bcc465afea9080 not found: ID does not exist" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.934825 4989 scope.go:117] "RemoveContainer" containerID="65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e" Oct 06 11:14:01 crc kubenswrapper[4989]: E1006 11:14:01.935327 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e\": container with ID starting with 65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e not found: ID does not exist" containerID="65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.935413 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e"} err="failed to get container status \"65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e\": rpc error: code = NotFound desc = could not find container \"65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e\": container with ID starting with 65f828e8f4820e95c0e26135c00664492999020eb032f0d1c5520470d666330e not found: ID does not exist" Oct 06 11:14:01 crc kubenswrapper[4989]: I1006 11:14:01.951498 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" path="/var/lib/kubelet/pods/244a8549-d6d2-4d9b-b1b8-0f089c2436c0/volumes" Oct 06 11:14:28 crc kubenswrapper[4989]: I1006 11:14:28.163724 4989 scope.go:117] "RemoveContainer" containerID="9d8088dc2276669ba6b3bd6460d5c18fcf6559d52cfcd7f4aa47c56a911288de" Oct 06 11:14:28 crc kubenswrapper[4989]: I1006 11:14:28.641770 4989 scope.go:117] "RemoveContainer" containerID="67bc295da46b455fae79eb069817722655b643277f36989f3f11c503b0d0aa12" Oct 06 11:14:28 crc kubenswrapper[4989]: I1006 11:14:28.694880 4989 scope.go:117] "RemoveContainer" containerID="0d629e2458e0dfb1cc7825ce525812276fa44d6e516f4ec916b1d2385eccc880" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.842313 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:14:55 crc kubenswrapper[4989]: E1006 11:14:55.847172 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="extract-content" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.847208 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="extract-content" Oct 06 11:14:55 crc kubenswrapper[4989]: E1006 11:14:55.847234 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="extract-utilities" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.847249 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="extract-utilities" Oct 06 11:14:55 crc kubenswrapper[4989]: E1006 11:14:55.847329 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="registry-server" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.847348 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="registry-server" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.847903 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="244a8549-d6d2-4d9b-b1b8-0f089c2436c0" containerName="registry-server" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.851991 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:55 crc kubenswrapper[4989]: I1006 11:14:55.895166 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.012802 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdzs7\" (UniqueName: \"kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.012895 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.012931 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.114660 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdzs7\" (UniqueName: \"kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.114753 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.114783 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.115414 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.115448 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.142361 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdzs7\" (UniqueName: \"kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7\") pod \"redhat-operators-49f2v\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.215994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:14:56 crc kubenswrapper[4989]: I1006 11:14:56.727232 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:14:57 crc kubenswrapper[4989]: I1006 11:14:57.612646 4989 generic.go:334] "Generic (PLEG): container finished" podID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerID="f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954" exitCode=0 Oct 06 11:14:57 crc kubenswrapper[4989]: I1006 11:14:57.612776 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerDied","Data":"f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954"} Oct 06 11:14:57 crc kubenswrapper[4989]: I1006 11:14:57.616095 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerStarted","Data":"7687edf1135fca4ef3bce477b5866645d3b2656f2b75a74ac00bcc42ad2154fd"} Oct 06 11:14:59 crc kubenswrapper[4989]: I1006 11:14:59.664342 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerStarted","Data":"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8"} Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.161275 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp"] Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.162954 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.165715 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.168140 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.182101 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp"] Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.242758 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84bzb\" (UniqueName: \"kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.243041 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.243177 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.344818 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.344929 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84bzb\" (UniqueName: \"kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.345033 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.346037 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.352359 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.364542 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84bzb\" (UniqueName: \"kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb\") pod \"collect-profiles-29329155-rsvcp\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:00 crc kubenswrapper[4989]: I1006 11:15:00.514404 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:01 crc kubenswrapper[4989]: I1006 11:15:01.052919 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp"] Oct 06 11:15:01 crc kubenswrapper[4989]: I1006 11:15:01.704765 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" event={"ID":"4700558f-a79f-4ada-8f7b-a5ebe0f11981","Type":"ContainerStarted","Data":"d09f706f90399ff97c5c12165b1ec37c1361c2c533937a61ace14246f3baeafa"} Oct 06 11:15:01 crc kubenswrapper[4989]: I1006 11:15:01.705247 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" event={"ID":"4700558f-a79f-4ada-8f7b-a5ebe0f11981","Type":"ContainerStarted","Data":"5b3179093fc73d91f9ef922fed9393a3f84ade5c88fdbcd54ec09a6b27aa4bc0"} Oct 06 11:15:01 crc kubenswrapper[4989]: I1006 11:15:01.733212 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" podStartSLOduration=1.7331931969999999 podStartE2EDuration="1.733193197s" podCreationTimestamp="2025-10-06 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:15:01.72702418 +0000 UTC m=+9352.517049770" watchObservedRunningTime="2025-10-06 11:15:01.733193197 +0000 UTC m=+9352.523218777" Oct 06 11:15:02 crc kubenswrapper[4989]: I1006 11:15:02.722668 4989 generic.go:334] "Generic (PLEG): container finished" podID="4700558f-a79f-4ada-8f7b-a5ebe0f11981" containerID="d09f706f90399ff97c5c12165b1ec37c1361c2c533937a61ace14246f3baeafa" exitCode=0 Oct 06 11:15:02 crc kubenswrapper[4989]: I1006 11:15:02.722851 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" event={"ID":"4700558f-a79f-4ada-8f7b-a5ebe0f11981","Type":"ContainerDied","Data":"d09f706f90399ff97c5c12165b1ec37c1361c2c533937a61ace14246f3baeafa"} Oct 06 11:15:02 crc kubenswrapper[4989]: I1006 11:15:02.726385 4989 generic.go:334] "Generic (PLEG): container finished" podID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerID="96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8" exitCode=0 Oct 06 11:15:02 crc kubenswrapper[4989]: I1006 11:15:02.726426 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerDied","Data":"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8"} Oct 06 11:15:03 crc kubenswrapper[4989]: I1006 11:15:03.746630 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerStarted","Data":"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1"} Oct 06 11:15:03 crc kubenswrapper[4989]: I1006 11:15:03.789365 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-49f2v" podStartSLOduration=2.984260043 podStartE2EDuration="8.789338738s" podCreationTimestamp="2025-10-06 11:14:55 +0000 UTC" firstStartedPulling="2025-10-06 11:14:57.615724926 +0000 UTC m=+9348.405750506" lastFinishedPulling="2025-10-06 11:15:03.420803581 +0000 UTC m=+9354.210829201" observedRunningTime="2025-10-06 11:15:03.768013404 +0000 UTC m=+9354.558039034" watchObservedRunningTime="2025-10-06 11:15:03.789338738 +0000 UTC m=+9354.579364358" Oct 06 11:15:03 crc kubenswrapper[4989]: I1006 11:15:03.934958 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:15:03 crc kubenswrapper[4989]: I1006 11:15:03.935195 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.140789 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.253474 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume\") pod \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.253551 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume\") pod \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.253642 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84bzb\" (UniqueName: \"kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb\") pod \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\" (UID: \"4700558f-a79f-4ada-8f7b-a5ebe0f11981\") " Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.254274 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume" (OuterVolumeSpecName: "config-volume") pod "4700558f-a79f-4ada-8f7b-a5ebe0f11981" (UID: "4700558f-a79f-4ada-8f7b-a5ebe0f11981"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.254739 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4700558f-a79f-4ada-8f7b-a5ebe0f11981-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.259795 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb" (OuterVolumeSpecName: "kube-api-access-84bzb") pod "4700558f-a79f-4ada-8f7b-a5ebe0f11981" (UID: "4700558f-a79f-4ada-8f7b-a5ebe0f11981"). InnerVolumeSpecName "kube-api-access-84bzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.269821 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4700558f-a79f-4ada-8f7b-a5ebe0f11981" (UID: "4700558f-a79f-4ada-8f7b-a5ebe0f11981"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.356533 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4700558f-a79f-4ada-8f7b-a5ebe0f11981-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.356582 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84bzb\" (UniqueName: \"kubernetes.io/projected/4700558f-a79f-4ada-8f7b-a5ebe0f11981-kube-api-access-84bzb\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.756263 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" event={"ID":"4700558f-a79f-4ada-8f7b-a5ebe0f11981","Type":"ContainerDied","Data":"5b3179093fc73d91f9ef922fed9393a3f84ade5c88fdbcd54ec09a6b27aa4bc0"} Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.756603 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b3179093fc73d91f9ef922fed9393a3f84ade5c88fdbcd54ec09a6b27aa4bc0" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.756315 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329155-rsvcp" Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.810309 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv"] Oct 06 11:15:04 crc kubenswrapper[4989]: I1006 11:15:04.819560 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329110-n6xqv"] Oct 06 11:15:05 crc kubenswrapper[4989]: I1006 11:15:05.950043 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="841bea16-2f0a-4327-b3b9-44c16d73213a" path="/var/lib/kubelet/pods/841bea16-2f0a-4327-b3b9-44c16d73213a/volumes" Oct 06 11:15:06 crc kubenswrapper[4989]: I1006 11:15:06.221141 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:06 crc kubenswrapper[4989]: I1006 11:15:06.221901 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:07 crc kubenswrapper[4989]: I1006 11:15:07.280739 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-49f2v" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="registry-server" probeResult="failure" output=< Oct 06 11:15:07 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 11:15:07 crc kubenswrapper[4989]: > Oct 06 11:15:16 crc kubenswrapper[4989]: I1006 11:15:16.315770 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:16 crc kubenswrapper[4989]: I1006 11:15:16.417814 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:16 crc kubenswrapper[4989]: I1006 11:15:16.590333 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:15:17 crc kubenswrapper[4989]: I1006 11:15:17.932164 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-49f2v" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="registry-server" containerID="cri-o://5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1" gracePeriod=2 Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.431940 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.577102 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities\") pod \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.577260 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content\") pod \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.577564 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdzs7\" (UniqueName: \"kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7\") pod \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\" (UID: \"7a44a610-5edb-41e6-b84a-e24eec5a3cec\") " Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.578251 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities" (OuterVolumeSpecName: "utilities") pod "7a44a610-5edb-41e6-b84a-e24eec5a3cec" (UID: "7a44a610-5edb-41e6-b84a-e24eec5a3cec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.578777 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.583912 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7" (OuterVolumeSpecName: "kube-api-access-zdzs7") pod "7a44a610-5edb-41e6-b84a-e24eec5a3cec" (UID: "7a44a610-5edb-41e6-b84a-e24eec5a3cec"). InnerVolumeSpecName "kube-api-access-zdzs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.660681 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a44a610-5edb-41e6-b84a-e24eec5a3cec" (UID: "7a44a610-5edb-41e6-b84a-e24eec5a3cec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.680641 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a44a610-5edb-41e6-b84a-e24eec5a3cec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.680685 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdzs7\" (UniqueName: \"kubernetes.io/projected/7a44a610-5edb-41e6-b84a-e24eec5a3cec-kube-api-access-zdzs7\") on node \"crc\" DevicePath \"\"" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.954841 4989 generic.go:334] "Generic (PLEG): container finished" podID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerID="5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1" exitCode=0 Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.954894 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerDied","Data":"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1"} Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.954924 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49f2v" event={"ID":"7a44a610-5edb-41e6-b84a-e24eec5a3cec","Type":"ContainerDied","Data":"7687edf1135fca4ef3bce477b5866645d3b2656f2b75a74ac00bcc42ad2154fd"} Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.954944 4989 scope.go:117] "RemoveContainer" containerID="5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1" Oct 06 11:15:18 crc kubenswrapper[4989]: I1006 11:15:18.955090 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49f2v" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.018016 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.023844 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-49f2v"] Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.028283 4989 scope.go:117] "RemoveContainer" containerID="96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.069464 4989 scope.go:117] "RemoveContainer" containerID="f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.133792 4989 scope.go:117] "RemoveContainer" containerID="5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1" Oct 06 11:15:19 crc kubenswrapper[4989]: E1006 11:15:19.134477 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1\": container with ID starting with 5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1 not found: ID does not exist" containerID="5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.134515 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1"} err="failed to get container status \"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1\": rpc error: code = NotFound desc = could not find container \"5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1\": container with ID starting with 5bbf3eb7687c6178ccc962bdbed7aa4f64e26fb311c3dfe3bb79694d450c72d1 not found: ID does not exist" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.134540 4989 scope.go:117] "RemoveContainer" containerID="96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8" Oct 06 11:15:19 crc kubenswrapper[4989]: E1006 11:15:19.134965 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8\": container with ID starting with 96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8 not found: ID does not exist" containerID="96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.135071 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8"} err="failed to get container status \"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8\": rpc error: code = NotFound desc = could not find container \"96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8\": container with ID starting with 96d263c7b9a338ed5e5addfbc60bd9d1d317ed35371525be45c9c3a6268302c8 not found: ID does not exist" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.135106 4989 scope.go:117] "RemoveContainer" containerID="f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954" Oct 06 11:15:19 crc kubenswrapper[4989]: E1006 11:15:19.135386 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954\": container with ID starting with f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954 not found: ID does not exist" containerID="f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.135413 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954"} err="failed to get container status \"f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954\": rpc error: code = NotFound desc = could not find container \"f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954\": container with ID starting with f1b5b7cc79cafc26dfdcab3953f365f7fc2c2edbebdfd465fdd3e01449683954 not found: ID does not exist" Oct 06 11:15:19 crc kubenswrapper[4989]: I1006 11:15:19.953549 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" path="/var/lib/kubelet/pods/7a44a610-5edb-41e6-b84a-e24eec5a3cec/volumes" Oct 06 11:15:28 crc kubenswrapper[4989]: I1006 11:15:28.786437 4989 scope.go:117] "RemoveContainer" containerID="42e08aa8e28abffdcc56ed8e4b6ad8d2837ce327ee55d8323aac87b9b82234c0" Oct 06 11:15:33 crc kubenswrapper[4989]: I1006 11:15:33.937948 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:15:33 crc kubenswrapper[4989]: I1006 11:15:33.940306 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:16:03 crc kubenswrapper[4989]: I1006 11:16:03.935785 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:16:03 crc kubenswrapper[4989]: I1006 11:16:03.936804 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:16:03 crc kubenswrapper[4989]: I1006 11:16:03.967942 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:16:03 crc kubenswrapper[4989]: I1006 11:16:03.969344 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:16:03 crc kubenswrapper[4989]: I1006 11:16:03.969468 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c" gracePeriod=600 Oct 06 11:16:04 crc kubenswrapper[4989]: I1006 11:16:04.596592 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c"} Oct 06 11:16:04 crc kubenswrapper[4989]: I1006 11:16:04.597358 4989 scope.go:117] "RemoveContainer" containerID="4ce70cc1a9beb9e735bf7640aade23d19baaf2a7c4a34b6737c5c900ac2b1e54" Oct 06 11:16:04 crc kubenswrapper[4989]: I1006 11:16:04.596546 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c" exitCode=0 Oct 06 11:16:04 crc kubenswrapper[4989]: I1006 11:16:04.597582 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84"} Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.667464 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:43 crc kubenswrapper[4989]: E1006 11:17:43.669354 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="registry-server" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669374 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="registry-server" Oct 06 11:17:43 crc kubenswrapper[4989]: E1006 11:17:43.669421 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="extract-utilities" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669433 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="extract-utilities" Oct 06 11:17:43 crc kubenswrapper[4989]: E1006 11:17:43.669474 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="extract-content" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669486 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="extract-content" Oct 06 11:17:43 crc kubenswrapper[4989]: E1006 11:17:43.669519 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4700558f-a79f-4ada-8f7b-a5ebe0f11981" containerName="collect-profiles" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669530 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="4700558f-a79f-4ada-8f7b-a5ebe0f11981" containerName="collect-profiles" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669868 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a44a610-5edb-41e6-b84a-e24eec5a3cec" containerName="registry-server" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.669905 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="4700558f-a79f-4ada-8f7b-a5ebe0f11981" containerName="collect-profiles" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.672826 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.693590 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.774824 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.774889 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.775300 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbp4\" (UniqueName: \"kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.877740 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.877997 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.878154 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.878166 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbp4\" (UniqueName: \"kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.878497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:43 crc kubenswrapper[4989]: I1006 11:17:43.899524 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbp4\" (UniqueName: \"kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4\") pod \"certified-operators-cwlp7\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:44 crc kubenswrapper[4989]: I1006 11:17:44.005197 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:44 crc kubenswrapper[4989]: I1006 11:17:44.537505 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:44 crc kubenswrapper[4989]: I1006 11:17:44.864525 4989 generic.go:334] "Generic (PLEG): container finished" podID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerID="75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb" exitCode=0 Oct 06 11:17:44 crc kubenswrapper[4989]: I1006 11:17:44.864594 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerDied","Data":"75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb"} Oct 06 11:17:44 crc kubenswrapper[4989]: I1006 11:17:44.864635 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerStarted","Data":"fe7a09cc5280e3b75a68e794469ddedcb784f27714422ba818f7e7fbcdc65562"} Oct 06 11:17:45 crc kubenswrapper[4989]: I1006 11:17:45.880349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerStarted","Data":"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f"} Oct 06 11:17:47 crc kubenswrapper[4989]: I1006 11:17:47.902069 4989 generic.go:334] "Generic (PLEG): container finished" podID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerID="bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f" exitCode=0 Oct 06 11:17:47 crc kubenswrapper[4989]: I1006 11:17:47.902161 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerDied","Data":"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f"} Oct 06 11:17:48 crc kubenswrapper[4989]: I1006 11:17:48.917692 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerStarted","Data":"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80"} Oct 06 11:17:48 crc kubenswrapper[4989]: I1006 11:17:48.952444 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cwlp7" podStartSLOduration=2.437957143 podStartE2EDuration="5.952418099s" podCreationTimestamp="2025-10-06 11:17:43 +0000 UTC" firstStartedPulling="2025-10-06 11:17:44.869463414 +0000 UTC m=+9515.659489004" lastFinishedPulling="2025-10-06 11:17:48.38392438 +0000 UTC m=+9519.173949960" observedRunningTime="2025-10-06 11:17:48.939544998 +0000 UTC m=+9519.729570608" watchObservedRunningTime="2025-10-06 11:17:48.952418099 +0000 UTC m=+9519.742443719" Oct 06 11:17:54 crc kubenswrapper[4989]: I1006 11:17:54.005888 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:54 crc kubenswrapper[4989]: I1006 11:17:54.006747 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:54 crc kubenswrapper[4989]: I1006 11:17:54.093923 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:55 crc kubenswrapper[4989]: I1006 11:17:55.035352 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:55 crc kubenswrapper[4989]: I1006 11:17:55.105529 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.017882 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cwlp7" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="registry-server" containerID="cri-o://313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80" gracePeriod=2 Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.618021 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.733624 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content\") pod \"caf30bc8-b23d-47f9-af0e-d3166170085a\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.733834 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbp4\" (UniqueName: \"kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4\") pod \"caf30bc8-b23d-47f9-af0e-d3166170085a\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.734151 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities\") pod \"caf30bc8-b23d-47f9-af0e-d3166170085a\" (UID: \"caf30bc8-b23d-47f9-af0e-d3166170085a\") " Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.737983 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities" (OuterVolumeSpecName: "utilities") pod "caf30bc8-b23d-47f9-af0e-d3166170085a" (UID: "caf30bc8-b23d-47f9-af0e-d3166170085a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.742513 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4" (OuterVolumeSpecName: "kube-api-access-4fbp4") pod "caf30bc8-b23d-47f9-af0e-d3166170085a" (UID: "caf30bc8-b23d-47f9-af0e-d3166170085a"). InnerVolumeSpecName "kube-api-access-4fbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.796859 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caf30bc8-b23d-47f9-af0e-d3166170085a" (UID: "caf30bc8-b23d-47f9-af0e-d3166170085a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.840404 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.840450 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caf30bc8-b23d-47f9-af0e-d3166170085a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:17:57 crc kubenswrapper[4989]: I1006 11:17:57.840467 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbp4\" (UniqueName: \"kubernetes.io/projected/caf30bc8-b23d-47f9-af0e-d3166170085a-kube-api-access-4fbp4\") on node \"crc\" DevicePath \"\"" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.030269 4989 generic.go:334] "Generic (PLEG): container finished" podID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerID="313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80" exitCode=0 Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.030324 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerDied","Data":"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80"} Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.030335 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cwlp7" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.030349 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cwlp7" event={"ID":"caf30bc8-b23d-47f9-af0e-d3166170085a","Type":"ContainerDied","Data":"fe7a09cc5280e3b75a68e794469ddedcb784f27714422ba818f7e7fbcdc65562"} Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.030365 4989 scope.go:117] "RemoveContainer" containerID="313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.062114 4989 scope.go:117] "RemoveContainer" containerID="bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.064057 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.077184 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cwlp7"] Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.092774 4989 scope.go:117] "RemoveContainer" containerID="75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.130033 4989 scope.go:117] "RemoveContainer" containerID="313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80" Oct 06 11:17:58 crc kubenswrapper[4989]: E1006 11:17:58.130472 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80\": container with ID starting with 313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80 not found: ID does not exist" containerID="313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.130517 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80"} err="failed to get container status \"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80\": rpc error: code = NotFound desc = could not find container \"313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80\": container with ID starting with 313c934144a6dd384d55e8060b158e3dad399614f7ae89c2ba4ae622a082ee80 not found: ID does not exist" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.130551 4989 scope.go:117] "RemoveContainer" containerID="bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f" Oct 06 11:17:58 crc kubenswrapper[4989]: E1006 11:17:58.130905 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f\": container with ID starting with bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f not found: ID does not exist" containerID="bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.130932 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f"} err="failed to get container status \"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f\": rpc error: code = NotFound desc = could not find container \"bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f\": container with ID starting with bf53187cc6ff1390b67882d0932c8a6bb116b8490d27ab9b8c4cb87bc4b7825f not found: ID does not exist" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.130943 4989 scope.go:117] "RemoveContainer" containerID="75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb" Oct 06 11:17:58 crc kubenswrapper[4989]: E1006 11:17:58.131507 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb\": container with ID starting with 75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb not found: ID does not exist" containerID="75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb" Oct 06 11:17:58 crc kubenswrapper[4989]: I1006 11:17:58.131549 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb"} err="failed to get container status \"75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb\": rpc error: code = NotFound desc = could not find container \"75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb\": container with ID starting with 75751d6a5047a56815d89a887c9fd48556282fcfad0de7f652d5a78a106b4afb not found: ID does not exist" Oct 06 11:17:59 crc kubenswrapper[4989]: I1006 11:17:59.952869 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" path="/var/lib/kubelet/pods/caf30bc8-b23d-47f9-af0e-d3166170085a/volumes" Oct 06 11:18:33 crc kubenswrapper[4989]: I1006 11:18:33.935830 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:18:33 crc kubenswrapper[4989]: I1006 11:18:33.936653 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:19:03 crc kubenswrapper[4989]: I1006 11:19:03.936517 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:19:03 crc kubenswrapper[4989]: I1006 11:19:03.937590 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:19:33 crc kubenswrapper[4989]: I1006 11:19:33.935597 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:19:33 crc kubenswrapper[4989]: I1006 11:19:33.936644 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:19:33 crc kubenswrapper[4989]: I1006 11:19:33.962719 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:19:33 crc kubenswrapper[4989]: I1006 11:19:33.964752 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:19:33 crc kubenswrapper[4989]: I1006 11:19:33.964903 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" gracePeriod=600 Oct 06 11:19:34 crc kubenswrapper[4989]: E1006 11:19:34.094942 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:19:34 crc kubenswrapper[4989]: E1006 11:19:34.219231 4989 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78ef9f2_a706_4a6c_8a9f_b232f97e9211.slice/crio-conmon-e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84.scope\": RecentStats: unable to find data in memory cache]" Oct 06 11:19:34 crc kubenswrapper[4989]: I1006 11:19:34.416465 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" exitCode=0 Oct 06 11:19:34 crc kubenswrapper[4989]: I1006 11:19:34.416541 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84"} Oct 06 11:19:34 crc kubenswrapper[4989]: I1006 11:19:34.416604 4989 scope.go:117] "RemoveContainer" containerID="6a71d4a20fd8d5b6849fc1894a753b70b69f13c4cae9266e8bf914c48f38902c" Oct 06 11:19:34 crc kubenswrapper[4989]: I1006 11:19:34.418433 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:19:34 crc kubenswrapper[4989]: E1006 11:19:34.421783 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:19:49 crc kubenswrapper[4989]: I1006 11:19:49.946448 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:19:49 crc kubenswrapper[4989]: E1006 11:19:49.947643 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:20:04 crc kubenswrapper[4989]: I1006 11:20:04.937733 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:20:04 crc kubenswrapper[4989]: E1006 11:20:04.939202 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:20:16 crc kubenswrapper[4989]: I1006 11:20:16.939331 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:20:16 crc kubenswrapper[4989]: E1006 11:20:16.945458 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:20:27 crc kubenswrapper[4989]: I1006 11:20:27.179239 4989 generic.go:334] "Generic (PLEG): container finished" podID="3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" containerID="5bc08c291a71b0cc6cecbf6b898d85f37d2db2fabfedbcf8f2628ddc34448dd5" exitCode=0 Oct 06 11:20:27 crc kubenswrapper[4989]: I1006 11:20:27.179366 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" event={"ID":"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4","Type":"ContainerDied","Data":"5bc08c291a71b0cc6cecbf6b898d85f37d2db2fabfedbcf8f2628ddc34448dd5"} Oct 06 11:20:27 crc kubenswrapper[4989]: I1006 11:20:27.936829 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:20:27 crc kubenswrapper[4989]: E1006 11:20:27.937910 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.820952 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939470 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939563 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939588 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939673 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939719 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939772 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939837 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939976 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwspj\" (UniqueName: \"kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.939997 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.940066 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.940254 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory\") pod \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\" (UID: \"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4\") " Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.953071 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.953133 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj" (OuterVolumeSpecName: "kube-api-access-gwspj") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "kube-api-access-gwspj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.976254 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph" (OuterVolumeSpecName: "ceph") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.976895 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.978750 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.994336 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:28 crc kubenswrapper[4989]: I1006 11:20:28.997236 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.005855 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.006268 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.007148 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory" (OuterVolumeSpecName: "inventory") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.008419 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" (UID: "3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045237 4989 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ceph\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045280 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045297 4989 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045311 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwspj\" (UniqueName: \"kubernetes.io/projected/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-kube-api-access-gwspj\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045325 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045339 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045353 4989 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-inventory\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045368 4989 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045383 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045394 4989 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.045408 4989 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.211495 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" event={"ID":"3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4","Type":"ContainerDied","Data":"1962ff8fd838da3d54fd9f7c35699f137f3e430cdc90b629ff47f877e5d6643c"} Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.211587 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1962ff8fd838da3d54fd9f7c35699f137f3e430cdc90b629ff47f877e5d6643c" Oct 06 11:20:29 crc kubenswrapper[4989]: I1006 11:20:29.211631 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw" Oct 06 11:20:40 crc kubenswrapper[4989]: I1006 11:20:40.936736 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:20:40 crc kubenswrapper[4989]: E1006 11:20:40.938278 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:20:55 crc kubenswrapper[4989]: I1006 11:20:55.938576 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:20:55 crc kubenswrapper[4989]: E1006 11:20:55.940024 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:21:03 crc kubenswrapper[4989]: E1006 11:21:03.485040 4989 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.23:35902->38.129.56.23:40321: write tcp 38.129.56.23:35902->38.129.56.23:40321: write: broken pipe Oct 06 11:21:08 crc kubenswrapper[4989]: I1006 11:21:08.937266 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:21:08 crc kubenswrapper[4989]: E1006 11:21:08.938580 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:21:19 crc kubenswrapper[4989]: I1006 11:21:19.950185 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:21:19 crc kubenswrapper[4989]: E1006 11:21:19.952507 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:21:34 crc kubenswrapper[4989]: I1006 11:21:34.937510 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:21:34 crc kubenswrapper[4989]: E1006 11:21:34.938587 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:21:48 crc kubenswrapper[4989]: I1006 11:21:48.937584 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:21:48 crc kubenswrapper[4989]: E1006 11:21:48.940007 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:22:03 crc kubenswrapper[4989]: I1006 11:22:03.936180 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:22:03 crc kubenswrapper[4989]: E1006 11:22:03.937198 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:22:14 crc kubenswrapper[4989]: I1006 11:22:14.937002 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:22:14 crc kubenswrapper[4989]: E1006 11:22:14.938559 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:22:29 crc kubenswrapper[4989]: I1006 11:22:29.956189 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:22:29 crc kubenswrapper[4989]: E1006 11:22:29.957724 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:22:41 crc kubenswrapper[4989]: I1006 11:22:41.937125 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:22:41 crc kubenswrapper[4989]: E1006 11:22:41.938333 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:22:53 crc kubenswrapper[4989]: I1006 11:22:53.686312 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 11:22:53 crc kubenswrapper[4989]: I1006 11:22:53.687335 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" containerName="adoption" containerID="cri-o://f8c31a9b0f716d1fa7864879dbc8d9457f62b077bf9eafb9939aedc1d96f90e0" gracePeriod=30 Oct 06 11:22:56 crc kubenswrapper[4989]: I1006 11:22:56.937189 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:22:56 crc kubenswrapper[4989]: E1006 11:22:56.938849 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:23:11 crc kubenswrapper[4989]: I1006 11:23:11.936710 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:23:11 crc kubenswrapper[4989]: E1006 11:23:11.937781 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.009250 4989 generic.go:334] "Generic (PLEG): container finished" podID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" containerID="f8c31a9b0f716d1fa7864879dbc8d9457f62b077bf9eafb9939aedc1d96f90e0" exitCode=137 Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.009328 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3","Type":"ContainerDied","Data":"f8c31a9b0f716d1fa7864879dbc8d9457f62b077bf9eafb9939aedc1d96f90e0"} Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.379887 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.419009 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb4sc\" (UniqueName: \"kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc\") pod \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.420437 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") pod \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\" (UID: \"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3\") " Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.430216 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc" (OuterVolumeSpecName: "kube-api-access-kb4sc") pod "204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" (UID: "204a66a8-2c30-4ce7-aed9-7fd6dfab01e3"). InnerVolumeSpecName "kube-api-access-kb4sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.450500 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25" (OuterVolumeSpecName: "mariadb-data") pod "204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" (UID: "204a66a8-2c30-4ce7-aed9-7fd6dfab01e3"). InnerVolumeSpecName "pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.523225 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb4sc\" (UniqueName: \"kubernetes.io/projected/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3-kube-api-access-kb4sc\") on node \"crc\" DevicePath \"\"" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.523293 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") on node \"crc\" " Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.554139 4989 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.554520 4989 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25") on node "crc" Oct 06 11:23:24 crc kubenswrapper[4989]: I1006 11:23:24.625597 4989 reconciler_common.go:293] "Volume detached for volume \"pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad66426d-e72d-46a3-b9d3-e79088a3ff25\") on node \"crc\" DevicePath \"\"" Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.025494 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"204a66a8-2c30-4ce7-aed9-7fd6dfab01e3","Type":"ContainerDied","Data":"a007c5219903258b6763b630ab901d11bb0c315ba99c9491548baadfb59dc164"} Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.026757 4989 scope.go:117] "RemoveContainer" containerID="f8c31a9b0f716d1fa7864879dbc8d9457f62b077bf9eafb9939aedc1d96f90e0" Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.025694 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.093626 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.101783 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.892157 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.892818 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="30510729-0329-47b2-ba21-8133a7559f05" containerName="adoption" containerID="cri-o://cd30673b163ccb8c4145a865b292e380edf28f5a7a31924e7dff26929382a556" gracePeriod=30 Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.936386 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:23:25 crc kubenswrapper[4989]: E1006 11:23:25.936892 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:23:25 crc kubenswrapper[4989]: I1006 11:23:25.950319 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" path="/var/lib/kubelet/pods/204a66a8-2c30-4ce7-aed9-7fd6dfab01e3/volumes" Oct 06 11:23:37 crc kubenswrapper[4989]: I1006 11:23:37.937859 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:23:37 crc kubenswrapper[4989]: E1006 11:23:37.940237 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.920870 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:23:47 crc kubenswrapper[4989]: E1006 11:23:47.923963 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.924086 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 06 11:23:47 crc kubenswrapper[4989]: E1006 11:23:47.924197 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="extract-utilities" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.924287 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="extract-utilities" Oct 06 11:23:47 crc kubenswrapper[4989]: E1006 11:23:47.924373 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="extract-content" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.924446 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="extract-content" Oct 06 11:23:47 crc kubenswrapper[4989]: E1006 11:23:47.924537 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="registry-server" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.924615 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="registry-server" Oct 06 11:23:47 crc kubenswrapper[4989]: E1006 11:23:47.924757 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" containerName="adoption" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.924846 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" containerName="adoption" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.925293 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.925456 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="204a66a8-2c30-4ce7-aed9-7fd6dfab01e3" containerName="adoption" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.925579 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf30bc8-b23d-47f9-af0e-d3166170085a" containerName="registry-server" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.928063 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:47 crc kubenswrapper[4989]: I1006 11:23:47.956433 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.064487 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hqcg\" (UniqueName: \"kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.064695 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.064776 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.166860 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.167109 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hqcg\" (UniqueName: \"kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.167207 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.167603 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.167603 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.193361 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hqcg\" (UniqueName: \"kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg\") pod \"redhat-marketplace-bd4dq\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.265807 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:48 crc kubenswrapper[4989]: I1006 11:23:48.830948 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:23:49 crc kubenswrapper[4989]: I1006 11:23:49.374944 4989 generic.go:334] "Generic (PLEG): container finished" podID="92322860-15a9-494e-b869-28fd7efe86bc" containerID="aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312" exitCode=0 Oct 06 11:23:49 crc kubenswrapper[4989]: I1006 11:23:49.375089 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerDied","Data":"aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312"} Oct 06 11:23:49 crc kubenswrapper[4989]: I1006 11:23:49.375530 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerStarted","Data":"53deb7bf39850ef307cbe0f971c2d11cea04d05e52769ffed50e033cb5c5f1f5"} Oct 06 11:23:49 crc kubenswrapper[4989]: I1006 11:23:49.383564 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:23:51 crc kubenswrapper[4989]: I1006 11:23:51.409788 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerStarted","Data":"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d"} Oct 06 11:23:51 crc kubenswrapper[4989]: I1006 11:23:51.937076 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:23:51 crc kubenswrapper[4989]: E1006 11:23:51.937698 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:23:52 crc kubenswrapper[4989]: I1006 11:23:52.432006 4989 generic.go:334] "Generic (PLEG): container finished" podID="92322860-15a9-494e-b869-28fd7efe86bc" containerID="18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d" exitCode=0 Oct 06 11:23:52 crc kubenswrapper[4989]: I1006 11:23:52.432090 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerDied","Data":"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d"} Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.280197 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.285941 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.297254 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.366671 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk2wf\" (UniqueName: \"kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.366775 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.366799 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.462140 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerStarted","Data":"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c"} Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.469393 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.469929 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.469994 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.470277 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.470521 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk2wf\" (UniqueName: \"kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.496839 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bd4dq" podStartSLOduration=3.475061397 podStartE2EDuration="7.49681743s" podCreationTimestamp="2025-10-06 11:23:47 +0000 UTC" firstStartedPulling="2025-10-06 11:23:49.382990432 +0000 UTC m=+9880.173016052" lastFinishedPulling="2025-10-06 11:23:53.404746485 +0000 UTC m=+9884.194772085" observedRunningTime="2025-10-06 11:23:54.496511751 +0000 UTC m=+9885.286537331" watchObservedRunningTime="2025-10-06 11:23:54.49681743 +0000 UTC m=+9885.286843010" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.497354 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk2wf\" (UniqueName: \"kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf\") pod \"community-operators-zrjp8\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:54 crc kubenswrapper[4989]: I1006 11:23:54.617273 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:23:55 crc kubenswrapper[4989]: I1006 11:23:55.218520 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.490249 4989 generic.go:334] "Generic (PLEG): container finished" podID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerID="4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca" exitCode=0 Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.490850 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerDied","Data":"4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca"} Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.490954 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerStarted","Data":"1dbfe39482d1f42e7fab9f85cfb85156f7cccedf77aba15beaf4616ccec01b34"} Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.503106 4989 generic.go:334] "Generic (PLEG): container finished" podID="30510729-0329-47b2-ba21-8133a7559f05" containerID="cd30673b163ccb8c4145a865b292e380edf28f5a7a31924e7dff26929382a556" exitCode=137 Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.507798 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"30510729-0329-47b2-ba21-8133a7559f05","Type":"ContainerDied","Data":"cd30673b163ccb8c4145a865b292e380edf28f5a7a31924e7dff26929382a556"} Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.581255 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.640859 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") pod \"30510729-0329-47b2-ba21-8133a7559f05\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.640955 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert\") pod \"30510729-0329-47b2-ba21-8133a7559f05\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.641102 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ltxj\" (UniqueName: \"kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj\") pod \"30510729-0329-47b2-ba21-8133a7559f05\" (UID: \"30510729-0329-47b2-ba21-8133a7559f05\") " Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.659970 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "30510729-0329-47b2-ba21-8133a7559f05" (UID: "30510729-0329-47b2-ba21-8133a7559f05"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.660538 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj" (OuterVolumeSpecName: "kube-api-access-2ltxj") pod "30510729-0329-47b2-ba21-8133a7559f05" (UID: "30510729-0329-47b2-ba21-8133a7559f05"). InnerVolumeSpecName "kube-api-access-2ltxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.672639 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e" (OuterVolumeSpecName: "ovn-data") pod "30510729-0329-47b2-ba21-8133a7559f05" (UID: "30510729-0329-47b2-ba21-8133a7559f05"). InnerVolumeSpecName "pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.745329 4989 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") on node \"crc\" " Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.745367 4989 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/30510729-0329-47b2-ba21-8133a7559f05-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.745379 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ltxj\" (UniqueName: \"kubernetes.io/projected/30510729-0329-47b2-ba21-8133a7559f05-kube-api-access-2ltxj\") on node \"crc\" DevicePath \"\"" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.777417 4989 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.777649 4989 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e") on node "crc" Oct 06 11:23:56 crc kubenswrapper[4989]: I1006 11:23:56.849327 4989 reconciler_common.go:293] "Volume detached for volume \"pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19b548bd-7ced-472d-be89-dcb4b4276e0e\") on node \"crc\" DevicePath \"\"" Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.521236 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerStarted","Data":"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9"} Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.523742 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"30510729-0329-47b2-ba21-8133a7559f05","Type":"ContainerDied","Data":"62311dd598feeac8d4ff0fef8990a84cbe3ba8efc84d5e36d4a51a8c05de7cac"} Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.523841 4989 scope.go:117] "RemoveContainer" containerID="cd30673b163ccb8c4145a865b292e380edf28f5a7a31924e7dff26929382a556" Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.523919 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.585098 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.597496 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 06 11:23:57 crc kubenswrapper[4989]: I1006 11:23:57.951979 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30510729-0329-47b2-ba21-8133a7559f05" path="/var/lib/kubelet/pods/30510729-0329-47b2-ba21-8133a7559f05/volumes" Oct 06 11:23:58 crc kubenswrapper[4989]: I1006 11:23:58.266876 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:58 crc kubenswrapper[4989]: I1006 11:23:58.266953 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:58 crc kubenswrapper[4989]: I1006 11:23:58.336786 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:58 crc kubenswrapper[4989]: I1006 11:23:58.632166 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:23:59 crc kubenswrapper[4989]: I1006 11:23:59.558751 4989 generic.go:334] "Generic (PLEG): container finished" podID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerID="39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9" exitCode=0 Oct 06 11:23:59 crc kubenswrapper[4989]: I1006 11:23:59.558836 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerDied","Data":"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9"} Oct 06 11:24:00 crc kubenswrapper[4989]: I1006 11:24:00.274915 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:24:00 crc kubenswrapper[4989]: I1006 11:24:00.576257 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerStarted","Data":"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab"} Oct 06 11:24:00 crc kubenswrapper[4989]: I1006 11:24:00.576511 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bd4dq" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="registry-server" containerID="cri-o://fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c" gracePeriod=2 Oct 06 11:24:00 crc kubenswrapper[4989]: I1006 11:24:00.630502 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zrjp8" podStartSLOduration=2.93924101 podStartE2EDuration="6.630471501s" podCreationTimestamp="2025-10-06 11:23:54 +0000 UTC" firstStartedPulling="2025-10-06 11:23:56.502810997 +0000 UTC m=+9887.292836577" lastFinishedPulling="2025-10-06 11:24:00.194041448 +0000 UTC m=+9890.984067068" observedRunningTime="2025-10-06 11:24:00.626101696 +0000 UTC m=+9891.416127326" watchObservedRunningTime="2025-10-06 11:24:00.630471501 +0000 UTC m=+9891.420497091" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.186646 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.305233 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities\") pod \"92322860-15a9-494e-b869-28fd7efe86bc\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.305616 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hqcg\" (UniqueName: \"kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg\") pod \"92322860-15a9-494e-b869-28fd7efe86bc\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.305831 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content\") pod \"92322860-15a9-494e-b869-28fd7efe86bc\" (UID: \"92322860-15a9-494e-b869-28fd7efe86bc\") " Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.309417 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities" (OuterVolumeSpecName: "utilities") pod "92322860-15a9-494e-b869-28fd7efe86bc" (UID: "92322860-15a9-494e-b869-28fd7efe86bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.314047 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg" (OuterVolumeSpecName: "kube-api-access-4hqcg") pod "92322860-15a9-494e-b869-28fd7efe86bc" (UID: "92322860-15a9-494e-b869-28fd7efe86bc"). InnerVolumeSpecName "kube-api-access-4hqcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.320317 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92322860-15a9-494e-b869-28fd7efe86bc" (UID: "92322860-15a9-494e-b869-28fd7efe86bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.409349 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hqcg\" (UniqueName: \"kubernetes.io/projected/92322860-15a9-494e-b869-28fd7efe86bc-kube-api-access-4hqcg\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.409385 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.409397 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92322860-15a9-494e-b869-28fd7efe86bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.595322 4989 generic.go:334] "Generic (PLEG): container finished" podID="92322860-15a9-494e-b869-28fd7efe86bc" containerID="fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c" exitCode=0 Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.595899 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerDied","Data":"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c"} Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.595972 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bd4dq" event={"ID":"92322860-15a9-494e-b869-28fd7efe86bc","Type":"ContainerDied","Data":"53deb7bf39850ef307cbe0f971c2d11cea04d05e52769ffed50e033cb5c5f1f5"} Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.596005 4989 scope.go:117] "RemoveContainer" containerID="fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.596268 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bd4dq" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.642222 4989 scope.go:117] "RemoveContainer" containerID="18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.675144 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.687820 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bd4dq"] Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.693707 4989 scope.go:117] "RemoveContainer" containerID="aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.770280 4989 scope.go:117] "RemoveContainer" containerID="fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c" Oct 06 11:24:01 crc kubenswrapper[4989]: E1006 11:24:01.770849 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c\": container with ID starting with fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c not found: ID does not exist" containerID="fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.770927 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c"} err="failed to get container status \"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c\": rpc error: code = NotFound desc = could not find container \"fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c\": container with ID starting with fb0f8b133254869c55edb71604f1440fc538d237aa60ee3ca69a9d9f2670515c not found: ID does not exist" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.770977 4989 scope.go:117] "RemoveContainer" containerID="18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d" Oct 06 11:24:01 crc kubenswrapper[4989]: E1006 11:24:01.771535 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d\": container with ID starting with 18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d not found: ID does not exist" containerID="18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.771580 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d"} err="failed to get container status \"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d\": rpc error: code = NotFound desc = could not find container \"18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d\": container with ID starting with 18410dec9928dc85fac7dfb864004ff5e3d8c578054ba7be5adbe9916089b91d not found: ID does not exist" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.771610 4989 scope.go:117] "RemoveContainer" containerID="aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312" Oct 06 11:24:01 crc kubenswrapper[4989]: E1006 11:24:01.772009 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312\": container with ID starting with aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312 not found: ID does not exist" containerID="aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.772061 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312"} err="failed to get container status \"aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312\": rpc error: code = NotFound desc = could not find container \"aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312\": container with ID starting with aa11d922c4a94de2fd12ab6c203a6690230f18a8bfc916ce9a4906140cee0312 not found: ID does not exist" Oct 06 11:24:01 crc kubenswrapper[4989]: I1006 11:24:01.952000 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92322860-15a9-494e-b869-28fd7efe86bc" path="/var/lib/kubelet/pods/92322860-15a9-494e-b869-28fd7efe86bc/volumes" Oct 06 11:24:02 crc kubenswrapper[4989]: I1006 11:24:02.936797 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:24:02 crc kubenswrapper[4989]: E1006 11:24:02.937854 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:24:04 crc kubenswrapper[4989]: I1006 11:24:04.618020 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:04 crc kubenswrapper[4989]: I1006 11:24:04.618496 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:04 crc kubenswrapper[4989]: I1006 11:24:04.717301 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:14 crc kubenswrapper[4989]: I1006 11:24:14.711649 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:14 crc kubenswrapper[4989]: I1006 11:24:14.809893 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:24:14 crc kubenswrapper[4989]: I1006 11:24:14.810949 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zrjp8" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="registry-server" containerID="cri-o://c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab" gracePeriod=2 Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.424897 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.489581 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content\") pod \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.490003 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities\") pod \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.490284 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk2wf\" (UniqueName: \"kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf\") pod \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\" (UID: \"7e04e5c2-9a70-4293-bb02-f1d146ec25be\") " Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.491635 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities" (OuterVolumeSpecName: "utilities") pod "7e04e5c2-9a70-4293-bb02-f1d146ec25be" (UID: "7e04e5c2-9a70-4293-bb02-f1d146ec25be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.507522 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf" (OuterVolumeSpecName: "kube-api-access-xk2wf") pod "7e04e5c2-9a70-4293-bb02-f1d146ec25be" (UID: "7e04e5c2-9a70-4293-bb02-f1d146ec25be"). InnerVolumeSpecName "kube-api-access-xk2wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.550615 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e04e5c2-9a70-4293-bb02-f1d146ec25be" (UID: "7e04e5c2-9a70-4293-bb02-f1d146ec25be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.594763 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk2wf\" (UniqueName: \"kubernetes.io/projected/7e04e5c2-9a70-4293-bb02-f1d146ec25be-kube-api-access-xk2wf\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.594808 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.594820 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e04e5c2-9a70-4293-bb02-f1d146ec25be-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.821339 4989 generic.go:334] "Generic (PLEG): container finished" podID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerID="c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab" exitCode=0 Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.821436 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerDied","Data":"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab"} Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.821545 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrjp8" event={"ID":"7e04e5c2-9a70-4293-bb02-f1d146ec25be","Type":"ContainerDied","Data":"1dbfe39482d1f42e7fab9f85cfb85156f7cccedf77aba15beaf4616ccec01b34"} Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.821479 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrjp8" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.821581 4989 scope.go:117] "RemoveContainer" containerID="c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.850578 4989 scope.go:117] "RemoveContainer" containerID="39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.889710 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.892387 4989 scope.go:117] "RemoveContainer" containerID="4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.903889 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zrjp8"] Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.937273 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:24:15 crc kubenswrapper[4989]: E1006 11:24:15.937920 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.955624 4989 scope.go:117] "RemoveContainer" containerID="c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab" Oct 06 11:24:15 crc kubenswrapper[4989]: E1006 11:24:15.956154 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab\": container with ID starting with c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab not found: ID does not exist" containerID="c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.956221 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab"} err="failed to get container status \"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab\": rpc error: code = NotFound desc = could not find container \"c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab\": container with ID starting with c558a4ba1b9a58eaacdae164a3efc105d1ac2be440a3ea7f03e6d1c5e95c60ab not found: ID does not exist" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.956300 4989 scope.go:117] "RemoveContainer" containerID="39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9" Oct 06 11:24:15 crc kubenswrapper[4989]: E1006 11:24:15.956719 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9\": container with ID starting with 39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9 not found: ID does not exist" containerID="39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.956761 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9"} err="failed to get container status \"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9\": rpc error: code = NotFound desc = could not find container \"39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9\": container with ID starting with 39d23759718e7ef4a86698a5c5fa3961d538a892d695f0019326b6207e86e7e9 not found: ID does not exist" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.956786 4989 scope.go:117] "RemoveContainer" containerID="4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.957560 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" path="/var/lib/kubelet/pods/7e04e5c2-9a70-4293-bb02-f1d146ec25be/volumes" Oct 06 11:24:15 crc kubenswrapper[4989]: E1006 11:24:15.957647 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca\": container with ID starting with 4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca not found: ID does not exist" containerID="4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca" Oct 06 11:24:15 crc kubenswrapper[4989]: I1006 11:24:15.957743 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca"} err="failed to get container status \"4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca\": rpc error: code = NotFound desc = could not find container \"4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca\": container with ID starting with 4a4edd89f5a09837283038bad7538514a6c78423d382c45b62d50ea61ac8bcca not found: ID does not exist" Oct 06 11:24:26 crc kubenswrapper[4989]: I1006 11:24:26.936066 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:24:26 crc kubenswrapper[4989]: E1006 11:24:26.937680 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:24:37 crc kubenswrapper[4989]: I1006 11:24:37.936481 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:24:39 crc kubenswrapper[4989]: I1006 11:24:39.223280 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052"} Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.725097 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvnb8/must-gather-rjtdk"] Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726433 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="extract-content" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726449 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="extract-content" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726467 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726474 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726499 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="extract-utilities" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726506 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="extract-utilities" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726527 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30510729-0329-47b2-ba21-8133a7559f05" containerName="adoption" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726536 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="30510729-0329-47b2-ba21-8133a7559f05" containerName="adoption" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726561 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726567 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726576 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="extract-utilities" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726584 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="extract-utilities" Oct 06 11:25:08 crc kubenswrapper[4989]: E1006 11:25:08.726601 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="extract-content" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726607 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="extract-content" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726879 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="92322860-15a9-494e-b869-28fd7efe86bc" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726912 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e04e5c2-9a70-4293-bb02-f1d146ec25be" containerName="registry-server" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.726922 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="30510729-0329-47b2-ba21-8133a7559f05" containerName="adoption" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.728227 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.732760 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qvnb8"/"openshift-service-ca.crt" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.757623 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvnb8/must-gather-rjtdk"] Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.758087 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qvnb8"/"kube-root-ca.crt" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.878438 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxbfn\" (UniqueName: \"kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.878878 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.982401 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxbfn\" (UniqueName: \"kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.982576 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:08 crc kubenswrapper[4989]: I1006 11:25:08.983355 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:09 crc kubenswrapper[4989]: I1006 11:25:09.003136 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxbfn\" (UniqueName: \"kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn\") pod \"must-gather-rjtdk\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:09 crc kubenswrapper[4989]: I1006 11:25:09.064123 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:25:09 crc kubenswrapper[4989]: I1006 11:25:09.851000 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qvnb8/must-gather-rjtdk"] Oct 06 11:25:10 crc kubenswrapper[4989]: I1006 11:25:10.708646 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" event={"ID":"bd58a75a-cc5b-489b-ac94-57daa6f651ae","Type":"ContainerStarted","Data":"df3261174c18a7fb17a4fd11436c1c0d0e653fd84891a898901442e247fb1ddd"} Oct 06 11:25:16 crc kubenswrapper[4989]: I1006 11:25:16.804259 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" event={"ID":"bd58a75a-cc5b-489b-ac94-57daa6f651ae","Type":"ContainerStarted","Data":"73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace"} Oct 06 11:25:16 crc kubenswrapper[4989]: I1006 11:25:16.805006 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" event={"ID":"bd58a75a-cc5b-489b-ac94-57daa6f651ae","Type":"ContainerStarted","Data":"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4"} Oct 06 11:25:16 crc kubenswrapper[4989]: I1006 11:25:16.845487 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" podStartSLOduration=3.210422166 podStartE2EDuration="8.845458843s" podCreationTimestamp="2025-10-06 11:25:08 +0000 UTC" firstStartedPulling="2025-10-06 11:25:09.863757437 +0000 UTC m=+9960.653783027" lastFinishedPulling="2025-10-06 11:25:15.498794124 +0000 UTC m=+9966.288819704" observedRunningTime="2025-10-06 11:25:16.827377622 +0000 UTC m=+9967.617403202" watchObservedRunningTime="2025-10-06 11:25:16.845458843 +0000 UTC m=+9967.635484453" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.095616 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rfz2j"] Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.098738 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.101045 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qvnb8"/"default-dockercfg-2n8kk" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.204683 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.204854 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnlzb\" (UniqueName: \"kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.307484 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.307577 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnlzb\" (UniqueName: \"kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.308193 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.328014 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnlzb\" (UniqueName: \"kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb\") pod \"crc-debug-rfz2j\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.426049 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:25:22 crc kubenswrapper[4989]: I1006 11:25:22.933851 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" event={"ID":"efcb7c88-0b73-49a4-8b6d-67f67b2e25da","Type":"ContainerStarted","Data":"c6229e3e22b0398d5324171ede81695a6457720d432e61d288e784a07adabf0c"} Oct 06 11:25:35 crc kubenswrapper[4989]: I1006 11:25:35.105890 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" event={"ID":"efcb7c88-0b73-49a4-8b6d-67f67b2e25da","Type":"ContainerStarted","Data":"559f83e9826e64082e2b36bac21f6d13d22fad55dc0cd6585609126a645c8a8e"} Oct 06 11:25:35 crc kubenswrapper[4989]: I1006 11:25:35.131481 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" podStartSLOduration=0.818804267 podStartE2EDuration="13.131460416s" podCreationTimestamp="2025-10-06 11:25:22 +0000 UTC" firstStartedPulling="2025-10-06 11:25:22.473339383 +0000 UTC m=+9973.263364963" lastFinishedPulling="2025-10-06 11:25:34.785995542 +0000 UTC m=+9985.576021112" observedRunningTime="2025-10-06 11:25:35.119675526 +0000 UTC m=+9985.909701096" watchObservedRunningTime="2025-10-06 11:25:35.131460416 +0000 UTC m=+9985.921485996" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.376393 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.379595 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.392816 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.551008 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wgp9\" (UniqueName: \"kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.551181 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.551333 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.669728 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.669862 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wgp9\" (UniqueName: \"kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.669948 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.671619 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.672189 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:51 crc kubenswrapper[4989]: I1006 11:25:51.717733 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wgp9\" (UniqueName: \"kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9\") pod \"redhat-operators-x55j8\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:52 crc kubenswrapper[4989]: I1006 11:25:52.016206 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:25:52 crc kubenswrapper[4989]: I1006 11:25:52.594057 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:25:52 crc kubenswrapper[4989]: W1006 11:25:52.645003 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecd47adb_2113_4eb3_bf98_8f32d99b959e.slice/crio-ac96c9b53089b4713920f79e9247e8f0052f773160f7a75dab015496d528ed59 WatchSource:0}: Error finding container ac96c9b53089b4713920f79e9247e8f0052f773160f7a75dab015496d528ed59: Status 404 returned error can't find the container with id ac96c9b53089b4713920f79e9247e8f0052f773160f7a75dab015496d528ed59 Oct 06 11:25:53 crc kubenswrapper[4989]: I1006 11:25:53.346957 4989 generic.go:334] "Generic (PLEG): container finished" podID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerID="95b80fd875fcfa38b399797c77b8a6678683153d855d76b2afdadf9d3c49da27" exitCode=0 Oct 06 11:25:53 crc kubenswrapper[4989]: I1006 11:25:53.347184 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerDied","Data":"95b80fd875fcfa38b399797c77b8a6678683153d855d76b2afdadf9d3c49da27"} Oct 06 11:25:53 crc kubenswrapper[4989]: I1006 11:25:53.347339 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerStarted","Data":"ac96c9b53089b4713920f79e9247e8f0052f773160f7a75dab015496d528ed59"} Oct 06 11:25:55 crc kubenswrapper[4989]: I1006 11:25:55.395101 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerStarted","Data":"eb6fd5497c8b89692a3880a993a5d44e959ed408f6d561b52af66a145325269a"} Oct 06 11:26:00 crc kubenswrapper[4989]: I1006 11:26:00.462929 4989 generic.go:334] "Generic (PLEG): container finished" podID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerID="eb6fd5497c8b89692a3880a993a5d44e959ed408f6d561b52af66a145325269a" exitCode=0 Oct 06 11:26:00 crc kubenswrapper[4989]: I1006 11:26:00.463032 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerDied","Data":"eb6fd5497c8b89692a3880a993a5d44e959ed408f6d561b52af66a145325269a"} Oct 06 11:26:02 crc kubenswrapper[4989]: I1006 11:26:02.502677 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerStarted","Data":"c4baa59adda9293ea4c92a057b1eb8f2c756d37156c28859ecd251199c9b912f"} Oct 06 11:26:02 crc kubenswrapper[4989]: I1006 11:26:02.529283 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x55j8" podStartSLOduration=3.764770652 podStartE2EDuration="11.529255896s" podCreationTimestamp="2025-10-06 11:25:51 +0000 UTC" firstStartedPulling="2025-10-06 11:25:53.350086308 +0000 UTC m=+10004.140111888" lastFinishedPulling="2025-10-06 11:26:01.114571552 +0000 UTC m=+10011.904597132" observedRunningTime="2025-10-06 11:26:02.526379563 +0000 UTC m=+10013.316405153" watchObservedRunningTime="2025-10-06 11:26:02.529255896 +0000 UTC m=+10013.319281496" Oct 06 11:26:12 crc kubenswrapper[4989]: I1006 11:26:12.017198 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:12 crc kubenswrapper[4989]: I1006 11:26:12.018115 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:13 crc kubenswrapper[4989]: I1006 11:26:13.109259 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x55j8" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" probeResult="failure" output=< Oct 06 11:26:13 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 11:26:13 crc kubenswrapper[4989]: > Oct 06 11:26:23 crc kubenswrapper[4989]: I1006 11:26:23.068473 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x55j8" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" probeResult="failure" output=< Oct 06 11:26:23 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 11:26:23 crc kubenswrapper[4989]: > Oct 06 11:26:33 crc kubenswrapper[4989]: I1006 11:26:33.096185 4989 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x55j8" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" probeResult="failure" output=< Oct 06 11:26:33 crc kubenswrapper[4989]: timeout: failed to connect service ":50051" within 1s Oct 06 11:26:33 crc kubenswrapper[4989]: > Oct 06 11:26:42 crc kubenswrapper[4989]: I1006 11:26:42.089255 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:42 crc kubenswrapper[4989]: I1006 11:26:42.160852 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:42 crc kubenswrapper[4989]: I1006 11:26:42.346913 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:26:43 crc kubenswrapper[4989]: I1006 11:26:43.979575 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x55j8" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" containerID="cri-o://c4baa59adda9293ea4c92a057b1eb8f2c756d37156c28859ecd251199c9b912f" gracePeriod=2 Oct 06 11:26:44 crc kubenswrapper[4989]: I1006 11:26:44.994112 4989 generic.go:334] "Generic (PLEG): container finished" podID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerID="c4baa59adda9293ea4c92a057b1eb8f2c756d37156c28859ecd251199c9b912f" exitCode=0 Oct 06 11:26:44 crc kubenswrapper[4989]: I1006 11:26:44.994187 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerDied","Data":"c4baa59adda9293ea4c92a057b1eb8f2c756d37156c28859ecd251199c9b912f"} Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.424624 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.542134 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities\") pod \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.542721 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content\") pod \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.542748 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wgp9\" (UniqueName: \"kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9\") pod \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\" (UID: \"ecd47adb-2113-4eb3-bf98-8f32d99b959e\") " Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.542777 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities" (OuterVolumeSpecName: "utilities") pod "ecd47adb-2113-4eb3-bf98-8f32d99b959e" (UID: "ecd47adb-2113-4eb3-bf98-8f32d99b959e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.544542 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.562156 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9" (OuterVolumeSpecName: "kube-api-access-6wgp9") pod "ecd47adb-2113-4eb3-bf98-8f32d99b959e" (UID: "ecd47adb-2113-4eb3-bf98-8f32d99b959e"). InnerVolumeSpecName "kube-api-access-6wgp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.629425 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecd47adb-2113-4eb3-bf98-8f32d99b959e" (UID: "ecd47adb-2113-4eb3-bf98-8f32d99b959e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.646638 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecd47adb-2113-4eb3-bf98-8f32d99b959e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:26:45 crc kubenswrapper[4989]: I1006 11:26:45.646690 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wgp9\" (UniqueName: \"kubernetes.io/projected/ecd47adb-2113-4eb3-bf98-8f32d99b959e-kube-api-access-6wgp9\") on node \"crc\" DevicePath \"\"" Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.018878 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x55j8" event={"ID":"ecd47adb-2113-4eb3-bf98-8f32d99b959e","Type":"ContainerDied","Data":"ac96c9b53089b4713920f79e9247e8f0052f773160f7a75dab015496d528ed59"} Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.018946 4989 scope.go:117] "RemoveContainer" containerID="c4baa59adda9293ea4c92a057b1eb8f2c756d37156c28859ecd251199c9b912f" Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.019192 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x55j8" Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.055715 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.069868 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x55j8"] Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.091913 4989 scope.go:117] "RemoveContainer" containerID="eb6fd5497c8b89692a3880a993a5d44e959ed408f6d561b52af66a145325269a" Oct 06 11:26:46 crc kubenswrapper[4989]: I1006 11:26:46.119951 4989 scope.go:117] "RemoveContainer" containerID="95b80fd875fcfa38b399797c77b8a6678683153d855d76b2afdadf9d3c49da27" Oct 06 11:26:47 crc kubenswrapper[4989]: I1006 11:26:47.950474 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" path="/var/lib/kubelet/pods/ecd47adb-2113-4eb3-bf98-8f32d99b959e/volumes" Oct 06 11:27:03 crc kubenswrapper[4989]: I1006 11:27:03.935195 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:27:03 crc kubenswrapper[4989]: I1006 11:27:03.936009 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:27:05 crc kubenswrapper[4989]: I1006 11:27:05.351912 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e899a871-db7d-484d-88b6-8f13b07e565d/init-config-reloader/0.log" Oct 06 11:27:05 crc kubenswrapper[4989]: I1006 11:27:05.586710 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e899a871-db7d-484d-88b6-8f13b07e565d/init-config-reloader/0.log" Oct 06 11:27:05 crc kubenswrapper[4989]: I1006 11:27:05.806448 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e899a871-db7d-484d-88b6-8f13b07e565d/config-reloader/0.log" Oct 06 11:27:05 crc kubenswrapper[4989]: I1006 11:27:05.815230 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e899a871-db7d-484d-88b6-8f13b07e565d/alertmanager/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.077714 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b89c8d47-43dd-4d8a-b31f-a7e5655d0427/aodh-api/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.322198 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b89c8d47-43dd-4d8a-b31f-a7e5655d0427/aodh-listener/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.348167 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b89c8d47-43dd-4d8a-b31f-a7e5655d0427/aodh-evaluator/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.562949 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_b89c8d47-43dd-4d8a-b31f-a7e5655d0427/aodh-notifier/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.776103 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65988db468-g5rf5_10c3c6f7-811a-40c0-92c2-63d02b2ecd2b/barbican-api/0.log" Oct 06 11:27:06 crc kubenswrapper[4989]: I1006 11:27:06.841033 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-65988db468-g5rf5_10c3c6f7-811a-40c0-92c2-63d02b2ecd2b/barbican-api-log/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.067362 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6bf65bc98-6gs8k_08847650-e4e2-465b-ba93-bab730da6f00/barbican-keystone-listener/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.347215 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6bf65bc98-6gs8k_08847650-e4e2-465b-ba93-bab730da6f00/barbican-keystone-listener-log/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.519968 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7f4d8b446c-sdh4z_af19d793-7709-494b-a5de-6696bc6ab72e/barbican-worker/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.561431 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7f4d8b446c-sdh4z_af19d793-7709-494b-a5de-6696bc6ab72e/barbican-worker-log/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.805719 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-hdzjv_f088c54d-a955-4cb4-a3af-7c3d1df3c5f1/bootstrap-openstack-openstack-cell1/0.log" Oct 06 11:27:07 crc kubenswrapper[4989]: I1006 11:27:07.969591 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d3514350-1ff6-43b2-b91b-a3781cb5050d/ceilometer-central-agent/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.115582 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d3514350-1ff6-43b2-b91b-a3781cb5050d/ceilometer-notification-agent/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.144174 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d3514350-1ff6-43b2-b91b-a3781cb5050d/proxy-httpd/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.193996 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d3514350-1ff6-43b2-b91b-a3781cb5050d/sg-core/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.354011 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-qmlnc_5fba31cf-6e75-4ffa-85e5-7d789ad6f92d/ceph-client-openstack-openstack-cell1/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.602007 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10ad6b44-28dc-4c46-bb6f-f3c6b898e334/cinder-api-log/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.683594 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10ad6b44-28dc-4c46-bb6f-f3c6b898e334/cinder-api/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.989247 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_67780570-63ad-40e2-88c9-dc4bb2563723/probe/0.log" Oct 06 11:27:08 crc kubenswrapper[4989]: I1006 11:27:08.989293 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_67780570-63ad-40e2-88c9-dc4bb2563723/cinder-backup/0.log" Oct 06 11:27:09 crc kubenswrapper[4989]: I1006 11:27:09.872514 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9e92a003-5bf6-4ee8-993f-cc75ee8bd077/cinder-scheduler/0.log" Oct 06 11:27:09 crc kubenswrapper[4989]: I1006 11:27:09.897835 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9e92a003-5bf6-4ee8-993f-cc75ee8bd077/probe/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.114231 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0fbd8084-52d9-4126-a357-c6e07168a716/cinder-volume/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.210165 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0fbd8084-52d9-4126-a357-c6e07168a716/probe/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.418085 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-9c25w_bf37b711-adc9-4b80-b9fc-efc5dcced79a/configure-network-openstack-openstack-cell1/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.679477 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-2xlln_ee637d80-17d5-4a6c-9734-67f46cdc47f0/configure-os-openstack-openstack-cell1/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.720193 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7cdb957947-rbss9_b35e50fb-921f-4cbe-adc6-6205aa5150f4/init/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.888577 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7cdb957947-rbss9_b35e50fb-921f-4cbe-adc6-6205aa5150f4/init/0.log" Oct 06 11:27:10 crc kubenswrapper[4989]: I1006 11:27:10.972965 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7cdb957947-rbss9_b35e50fb-921f-4cbe-adc6-6205aa5150f4/dnsmasq-dns/0.log" Oct 06 11:27:11 crc kubenswrapper[4989]: I1006 11:27:11.116963 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-65277_8f67e476-79a2-499d-ad1b-142eac5e3681/download-cache-openstack-openstack-cell1/0.log" Oct 06 11:27:11 crc kubenswrapper[4989]: I1006 11:27:11.249755 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ed83ef55-63eb-4626-9661-d64e098a04db/glance-httpd/0.log" Oct 06 11:27:11 crc kubenswrapper[4989]: I1006 11:27:11.331873 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ed83ef55-63eb-4626-9661-d64e098a04db/glance-log/0.log" Oct 06 11:27:11 crc kubenswrapper[4989]: I1006 11:27:11.507921 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed/glance-httpd/0.log" Oct 06 11:27:11 crc kubenswrapper[4989]: I1006 11:27:11.550893 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a3c491b5-c32c-4e2e-bbf1-5ea7bd33e1ed/glance-log/0.log" Oct 06 11:27:12 crc kubenswrapper[4989]: I1006 11:27:12.573455 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-6fd69d8487-9nkb7_5d0a7503-18c3-4f61-b6e4-7a083fb7d5fb/heat-api/0.log" Oct 06 11:27:12 crc kubenswrapper[4989]: I1006 11:27:12.837151 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-76b6b48b9d-lr9tc_78fc292c-9988-4bd5-a61f-754743dc1729/heat-engine/0.log" Oct 06 11:27:12 crc kubenswrapper[4989]: I1006 11:27:12.898061 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5cf7458ccf-9rtfv_9e1a109a-be53-4da4-b947-cf4184f43a0b/heat-cfnapi/0.log" Oct 06 11:27:13 crc kubenswrapper[4989]: I1006 11:27:13.140703 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f868b9f5c-rq2jl_157d6322-0be6-423b-8276-c789369b07d2/horizon/0.log" Oct 06 11:27:13 crc kubenswrapper[4989]: I1006 11:27:13.195630 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f868b9f5c-rq2jl_157d6322-0be6-423b-8276-c789369b07d2/horizon-log/0.log" Oct 06 11:27:13 crc kubenswrapper[4989]: I1006 11:27:13.422368 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-lfx9n_d3cacb3e-da1d-4bb8-b60a-890f8a8f935a/install-certs-openstack-openstack-cell1/0.log" Oct 06 11:27:13 crc kubenswrapper[4989]: I1006 11:27:13.609453 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-gwb5g_5204bd93-cd20-44b3-9f50-8f4113d8ab62/install-os-openstack-openstack-cell1/0.log" Oct 06 11:27:13 crc kubenswrapper[4989]: I1006 11:27:13.900761 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29329141-9vw8m_a5a0b07a-b289-4a82-a61f-02b5dc8b2c3a/keystone-cron/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.019968 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-555c6df57c-dz9gw_32295746-8fcc-4a25-87e3-75ab110afd5a/keystone-api/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.134676 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6c6f6eab-ae0b-4cab-b028-b589004ef31a/kube-state-metrics/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.399742 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-kr879_348d43fe-9095-4a19-b9ac-b3dc9e5da427/libvirt-openstack-openstack-cell1/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.638380 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1d4280fd-c098-4b4b-a926-123012f6b048/manila-api-log/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.661007 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1d4280fd-c098-4b4b-a926-123012f6b048/manila-api/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.893405 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_53c8f6a6-d2c2-4e4f-abed-30eb20208a7a/probe/0.log" Oct 06 11:27:14 crc kubenswrapper[4989]: I1006 11:27:14.958405 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_53c8f6a6-d2c2-4e4f-abed-30eb20208a7a/manila-scheduler/0.log" Oct 06 11:27:15 crc kubenswrapper[4989]: I1006 11:27:15.502590 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_12cdf345-49c9-4a58-9357-7cdf2b7622eb/manila-share/0.log" Oct 06 11:27:15 crc kubenswrapper[4989]: I1006 11:27:15.554885 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_12cdf345-49c9-4a58-9357-7cdf2b7622eb/probe/0.log" Oct 06 11:27:16 crc kubenswrapper[4989]: I1006 11:27:16.076795 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7954959975-7hbqd_0d12ecc3-3794-407e-a6e7-38198dc82213/neutron-api/0.log" Oct 06 11:27:16 crc kubenswrapper[4989]: I1006 11:27:16.169136 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7954959975-7hbqd_0d12ecc3-3794-407e-a6e7-38198dc82213/neutron-httpd/0.log" Oct 06 11:27:16 crc kubenswrapper[4989]: I1006 11:27:16.567527 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-jhx2r_1f292596-66c1-485a-bd87-31d9836fcd48/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 06 11:27:16 crc kubenswrapper[4989]: I1006 11:27:16.877366 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-89hhj_1906849c-8f07-411b-b7dc-974e982b96f8/neutron-metadata-openstack-openstack-cell1/0.log" Oct 06 11:27:17 crc kubenswrapper[4989]: I1006 11:27:17.177562 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-5hmfm_d24e47c9-fab9-49dc-9fb3-728e9611a342/neutron-sriov-openstack-openstack-cell1/0.log" Oct 06 11:27:17 crc kubenswrapper[4989]: I1006 11:27:17.593983 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_430c6cf8-e65b-4c3e-aeec-6a22e2f472b5/nova-api-api/0.log" Oct 06 11:27:17 crc kubenswrapper[4989]: I1006 11:27:17.799983 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_430c6cf8-e65b-4c3e-aeec-6a22e2f472b5/nova-api-log/0.log" Oct 06 11:27:18 crc kubenswrapper[4989]: I1006 11:27:18.208212 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3a757cac-8dd8-42c8-9cdc-4cf3533b9f77/nova-cell0-conductor-conductor/0.log" Oct 06 11:27:18 crc kubenswrapper[4989]: I1006 11:27:18.519885 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_301be596-d22c-4d61-9436-7f93b25bbd8a/nova-cell1-conductor-conductor/0.log" Oct 06 11:27:18 crc kubenswrapper[4989]: I1006 11:27:18.555670 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_ea1fbbbe-057b-4682-8900-d5c17675a000/memcached/0.log" Oct 06 11:27:18 crc kubenswrapper[4989]: I1006 11:27:18.756830 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_82f58534-c92c-4367-a1ca-1dfcef75384b/nova-cell1-novncproxy-novncproxy/0.log" Oct 06 11:27:18 crc kubenswrapper[4989]: I1006 11:27:18.911926 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell6w4rw_3cc1d11b-da7d-4b21-aa8a-2b5ca853ace4/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 06 11:27:19 crc kubenswrapper[4989]: I1006 11:27:19.200223 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-7qfsv_3c116bb3-99b0-4a49-8591-40a42aa83a3d/nova-cell1-openstack-openstack-cell1/0.log" Oct 06 11:27:19 crc kubenswrapper[4989]: I1006 11:27:19.357818 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_198f79e4-9643-4782-b5b8-269d0b51d986/nova-metadata-log/0.log" Oct 06 11:27:19 crc kubenswrapper[4989]: I1006 11:27:19.468196 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_198f79e4-9643-4782-b5b8-269d0b51d986/nova-metadata-metadata/0.log" Oct 06 11:27:19 crc kubenswrapper[4989]: I1006 11:27:19.653001 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0cd872d1-fade-451e-82bc-14993f72da4a/nova-scheduler-scheduler/0.log" Oct 06 11:27:19 crc kubenswrapper[4989]: I1006 11:27:19.726296 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db77676cd-jw67c_dfbd9202-a3ec-478c-a15d-40927aa5cf9a/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.035473 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db77676cd-jw67c_dfbd9202-a3ec-478c-a15d-40927aa5cf9a/octavia-api-provider-agent/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.068963 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db77676cd-jw67c_dfbd9202-a3ec-478c-a15d-40927aa5cf9a/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.242689 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-db77676cd-jw67c_dfbd9202-a3ec-478c-a15d-40927aa5cf9a/octavia-api/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.354040 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-p2zxx_b3e3fae2-2565-4d6b-8a43-30e48c1f734c/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.498481 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-p2zxx_b3e3fae2-2565-4d6b-8a43-30e48c1f734c/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.552168 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-p2zxx_b3e3fae2-2565-4d6b-8a43-30e48c1f734c/octavia-healthmanager/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.697025 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-7j2m4_a2fcb796-037b-403e-a59b-2d73a6e94de7/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.923107 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-7j2m4_a2fcb796-037b-403e-a59b-2d73a6e94de7/init/0.log" Oct 06 11:27:20 crc kubenswrapper[4989]: I1006 11:27:20.962613 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-7j2m4_a2fcb796-037b-403e-a59b-2d73a6e94de7/octavia-housekeeping/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.202161 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-hfddc_70a61785-944c-4912-8373-2e5161470a83/init/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.310327 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-hfddc_70a61785-944c-4912-8373-2e5161470a83/octavia-amphora-httpd/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.362415 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-hfddc_70a61785-944c-4912-8373-2e5161470a83/init/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.569044 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-wsxs8_9294460e-ff38-4149-b42f-5ca293adb874/init/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.773899 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-wsxs8_9294460e-ff38-4149-b42f-5ca293adb874/init/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.829579 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-wsxs8_9294460e-ff38-4149-b42f-5ca293adb874/octavia-rsyslog/0.log" Oct 06 11:27:21 crc kubenswrapper[4989]: I1006 11:27:21.995075 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-hkh7r_e308652c-1db7-49f7-bc52-c0b1e7818138/init/0.log" Oct 06 11:27:22 crc kubenswrapper[4989]: I1006 11:27:22.440731 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-hkh7r_e308652c-1db7-49f7-bc52-c0b1e7818138/init/0.log" Oct 06 11:27:22 crc kubenswrapper[4989]: I1006 11:27:22.622087 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-hkh7r_e308652c-1db7-49f7-bc52-c0b1e7818138/octavia-worker/0.log" Oct 06 11:27:22 crc kubenswrapper[4989]: I1006 11:27:22.706518 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6b4c9853-f4c0-4e59-9d5a-78af6bf457b1/mysql-bootstrap/0.log" Oct 06 11:27:22 crc kubenswrapper[4989]: I1006 11:27:22.873406 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6b4c9853-f4c0-4e59-9d5a-78af6bf457b1/mysql-bootstrap/0.log" Oct 06 11:27:22 crc kubenswrapper[4989]: I1006 11:27:22.923436 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_6b4c9853-f4c0-4e59-9d5a-78af6bf457b1/galera/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.164924 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_678e9748-bfb5-47c1-8788-083b5189145c/mysql-bootstrap/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.319053 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_678e9748-bfb5-47c1-8788-083b5189145c/mysql-bootstrap/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.377353 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_678e9748-bfb5-47c1-8788-083b5189145c/galera/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.528991 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_5cc043d8-b621-49dd-a595-d704df100397/openstackclient/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.779107 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8r8dw_47747799-142b-42f8-ad3a-ba483ad279d1/ovn-controller/0.log" Oct 06 11:27:23 crc kubenswrapper[4989]: I1006 11:27:23.939726 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9lbv7_764a099f-2698-470d-a6bc-d46435ac4951/openstack-network-exporter/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.128800 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pkfng_858946e9-0129-43b4-9044-7878d1e0b562/ovsdb-server-init/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.321965 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pkfng_858946e9-0129-43b4-9044-7878d1e0b562/ovsdb-server-init/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.326922 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pkfng_858946e9-0129-43b4-9044-7878d1e0b562/ovs-vswitchd/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.334244 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-pkfng_858946e9-0129-43b4-9044-7878d1e0b562/ovsdb-server/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.547424 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6e62e912-c83a-4f76-904e-45d246480cc6/openstack-network-exporter/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.553558 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_6e62e912-c83a-4f76-904e-45d246480cc6/ovn-northd/0.log" Oct 06 11:27:24 crc kubenswrapper[4989]: I1006 11:27:24.884386 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-fw472_e83646ac-1df3-40bc-8706-53b5e1de675f/ovn-openstack-openstack-cell1/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.029752 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af55f2bd-22b6-4604-9d04-03dbe5aaf3a6/openstack-network-exporter/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.167732 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af55f2bd-22b6-4604-9d04-03dbe5aaf3a6/ovsdbserver-nb/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.245184 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_c049c234-54cd-4801-9e73-41c5cb8d1404/openstack-network-exporter/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.418172 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_c049c234-54cd-4801-9e73-41c5cb8d1404/ovsdbserver-nb/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.463195 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_0dabb4ae-66bb-4073-af87-e8a398b22b1a/openstack-network-exporter/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.691894 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_0dabb4ae-66bb-4073-af87-e8a398b22b1a/ovsdbserver-nb/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.750004 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_27ebef9a-25d5-437a-bd77-5b8ccd0110ba/openstack-network-exporter/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.893880 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_27ebef9a-25d5-437a-bd77-5b8ccd0110ba/ovsdbserver-sb/0.log" Oct 06 11:27:25 crc kubenswrapper[4989]: I1006 11:27:25.958547 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_bb75f7eb-e857-4e0c-8900-bd9fea672fb5/openstack-network-exporter/0.log" Oct 06 11:27:26 crc kubenswrapper[4989]: I1006 11:27:26.132192 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_bb75f7eb-e857-4e0c-8900-bd9fea672fb5/ovsdbserver-sb/0.log" Oct 06 11:27:26 crc kubenswrapper[4989]: I1006 11:27:26.306871 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_fb3fa8a8-b876-485e-954f-3ed1c711bd61/ovsdbserver-sb/0.log" Oct 06 11:27:26 crc kubenswrapper[4989]: I1006 11:27:26.308334 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_fb3fa8a8-b876-485e-954f-3ed1c711bd61/openstack-network-exporter/0.log" Oct 06 11:27:26 crc kubenswrapper[4989]: I1006 11:27:26.924372 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849964bbfd-whdvn_4ee9b21b-081f-494d-b8de-7f4919a2d16c/placement-api/0.log" Oct 06 11:27:26 crc kubenswrapper[4989]: I1006 11:27:26.928595 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-849964bbfd-whdvn_4ee9b21b-081f-494d-b8de-7f4919a2d16c/placement-log/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.106451 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-czb7wd_a78e83ee-accd-45a6-be14-5d73b7ce2060/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.294308 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f5788673-b226-4b7d-a99d-dea8d93d1bed/init-config-reloader/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.484309 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f5788673-b226-4b7d-a99d-dea8d93d1bed/config-reloader/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.507807 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f5788673-b226-4b7d-a99d-dea8d93d1bed/init-config-reloader/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.537277 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f5788673-b226-4b7d-a99d-dea8d93d1bed/prometheus/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.705890 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f5788673-b226-4b7d-a99d-dea8d93d1bed/thanos-sidecar/0.log" Oct 06 11:27:27 crc kubenswrapper[4989]: I1006 11:27:27.731297 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf/setup-container/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.037591 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf/setup-container/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.061819 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_4c0afc56-a62b-4f8f-b1c2-b4bc1085d1bf/rabbitmq/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.220260 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_df5fb509-bd0b-42b3-95d4-4a689e8713e5/setup-container/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.416198 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_df5fb509-bd0b-42b3-95d4-4a689e8713e5/setup-container/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.468624 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_df5fb509-bd0b-42b3-95d4-4a689e8713e5/rabbitmq/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.606260 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-8nj2v_d8632b29-9ab1-4376-8b89-6be9d15e7754/reboot-os-openstack-openstack-cell1/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.750918 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-kb9m6_889a4162-2caf-4c70-8b84-f406cfb67f4a/run-os-openstack-openstack-cell1/0.log" Oct 06 11:27:28 crc kubenswrapper[4989]: I1006 11:27:28.971742 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-r29s8_7eaa911b-7e42-410e-a0f6-36abf0909e7a/ssh-known-hosts-openstack/0.log" Oct 06 11:27:29 crc kubenswrapper[4989]: I1006 11:27:29.138722 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-4xbqs_68affe53-42a8-400e-824b-af80c5c9e0e4/telemetry-openstack-openstack-cell1/0.log" Oct 06 11:27:29 crc kubenswrapper[4989]: I1006 11:27:29.361678 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-zt8dw_5e33f7f9-1020-4783-bee8-c3effc5defb2/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 06 11:27:29 crc kubenswrapper[4989]: I1006 11:27:29.525114 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-5cjdc_338161e0-27f7-4a63-86c8-1e35e92a352c/validate-network-openstack-openstack-cell1/0.log" Oct 06 11:27:33 crc kubenswrapper[4989]: I1006 11:27:33.939275 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:27:33 crc kubenswrapper[4989]: I1006 11:27:33.940258 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.257593 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:27:43 crc kubenswrapper[4989]: E1006 11:27:43.259235 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="extract-content" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.259260 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="extract-content" Oct 06 11:27:43 crc kubenswrapper[4989]: E1006 11:27:43.259284 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="extract-utilities" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.259298 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="extract-utilities" Oct 06 11:27:43 crc kubenswrapper[4989]: E1006 11:27:43.259359 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.259374 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.259849 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd47adb-2113-4eb3-bf98-8f32d99b959e" containerName="registry-server" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.264076 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.276949 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.349298 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldbps\" (UniqueName: \"kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.349416 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.349454 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.451870 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldbps\" (UniqueName: \"kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.452022 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.452066 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.452758 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.452790 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.483875 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldbps\" (UniqueName: \"kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps\") pod \"certified-operators-jxtps\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:43 crc kubenswrapper[4989]: I1006 11:27:43.587994 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:44 crc kubenswrapper[4989]: I1006 11:27:44.338292 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:27:44 crc kubenswrapper[4989]: I1006 11:27:44.770318 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerID="a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb" exitCode=0 Oct 06 11:27:44 crc kubenswrapper[4989]: I1006 11:27:44.770790 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerDied","Data":"a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb"} Oct 06 11:27:44 crc kubenswrapper[4989]: I1006 11:27:44.770832 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerStarted","Data":"928ddec2788c547d9ffc14862e64f801d91e90c7f969bc891bd36ee2ccb4487a"} Oct 06 11:27:46 crc kubenswrapper[4989]: I1006 11:27:46.800474 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerStarted","Data":"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2"} Oct 06 11:27:48 crc kubenswrapper[4989]: I1006 11:27:48.827097 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerID="e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2" exitCode=0 Oct 06 11:27:48 crc kubenswrapper[4989]: I1006 11:27:48.827168 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerDied","Data":"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2"} Oct 06 11:27:49 crc kubenswrapper[4989]: I1006 11:27:49.845418 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerStarted","Data":"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104"} Oct 06 11:27:49 crc kubenswrapper[4989]: I1006 11:27:49.885182 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jxtps" podStartSLOduration=2.296361744 podStartE2EDuration="6.885129533s" podCreationTimestamp="2025-10-06 11:27:43 +0000 UTC" firstStartedPulling="2025-10-06 11:27:44.772823326 +0000 UTC m=+10115.562848896" lastFinishedPulling="2025-10-06 11:27:49.361591105 +0000 UTC m=+10120.151616685" observedRunningTime="2025-10-06 11:27:49.867113674 +0000 UTC m=+10120.657139254" watchObservedRunningTime="2025-10-06 11:27:49.885129533 +0000 UTC m=+10120.675155153" Oct 06 11:27:53 crc kubenswrapper[4989]: I1006 11:27:53.588806 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:53 crc kubenswrapper[4989]: I1006 11:27:53.589301 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:27:53 crc kubenswrapper[4989]: I1006 11:27:53.656903 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.648397 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.727521 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.936117 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.936221 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.965407 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.966472 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:28:03 crc kubenswrapper[4989]: I1006 11:28:03.966555 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052" gracePeriod=600 Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.033512 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jxtps" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="registry-server" containerID="cri-o://7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104" gracePeriod=2 Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.644077 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.714404 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities\") pod \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.714474 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldbps\" (UniqueName: \"kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps\") pod \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.714966 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content\") pod \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\" (UID: \"aba00a00-6135-4e37-b0a0-69fd11d45bbc\") " Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.715827 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities" (OuterVolumeSpecName: "utilities") pod "aba00a00-6135-4e37-b0a0-69fd11d45bbc" (UID: "aba00a00-6135-4e37-b0a0-69fd11d45bbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.723513 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps" (OuterVolumeSpecName: "kube-api-access-ldbps") pod "aba00a00-6135-4e37-b0a0-69fd11d45bbc" (UID: "aba00a00-6135-4e37-b0a0-69fd11d45bbc"). InnerVolumeSpecName "kube-api-access-ldbps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.765251 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aba00a00-6135-4e37-b0a0-69fd11d45bbc" (UID: "aba00a00-6135-4e37-b0a0-69fd11d45bbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.827894 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.827937 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aba00a00-6135-4e37-b0a0-69fd11d45bbc-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:04 crc kubenswrapper[4989]: I1006 11:28:04.827948 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldbps\" (UniqueName: \"kubernetes.io/projected/aba00a00-6135-4e37-b0a0-69fd11d45bbc-kube-api-access-ldbps\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.050950 4989 generic.go:334] "Generic (PLEG): container finished" podID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerID="7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104" exitCode=0 Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.051031 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerDied","Data":"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104"} Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.051070 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxtps" event={"ID":"aba00a00-6135-4e37-b0a0-69fd11d45bbc","Type":"ContainerDied","Data":"928ddec2788c547d9ffc14862e64f801d91e90c7f969bc891bd36ee2ccb4487a"} Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.051096 4989 scope.go:117] "RemoveContainer" containerID="7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.051319 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxtps" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.054447 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052" exitCode=0 Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.054521 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052"} Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.054576 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerStarted","Data":"5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36"} Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.144120 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.147359 4989 scope.go:117] "RemoveContainer" containerID="e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.213833 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jxtps"] Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.236802 4989 scope.go:117] "RemoveContainer" containerID="a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.264909 4989 scope.go:117] "RemoveContainer" containerID="7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104" Oct 06 11:28:05 crc kubenswrapper[4989]: E1006 11:28:05.265312 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104\": container with ID starting with 7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104 not found: ID does not exist" containerID="7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.265372 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104"} err="failed to get container status \"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104\": rpc error: code = NotFound desc = could not find container \"7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104\": container with ID starting with 7c150b0bec9497dc2fa1aa8037d6171d1b88925bc06be20f873fa20c5e671104 not found: ID does not exist" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.265399 4989 scope.go:117] "RemoveContainer" containerID="e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2" Oct 06 11:28:05 crc kubenswrapper[4989]: E1006 11:28:05.265716 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2\": container with ID starting with e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2 not found: ID does not exist" containerID="e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.265755 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2"} err="failed to get container status \"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2\": rpc error: code = NotFound desc = could not find container \"e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2\": container with ID starting with e8d3025f61988ef6e9032f85d1531bade7c2c8bff6bfc84dcdc513b0cb84c6c2 not found: ID does not exist" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.265811 4989 scope.go:117] "RemoveContainer" containerID="a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb" Oct 06 11:28:05 crc kubenswrapper[4989]: E1006 11:28:05.266422 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb\": container with ID starting with a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb not found: ID does not exist" containerID="a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.266444 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb"} err="failed to get container status \"a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb\": rpc error: code = NotFound desc = could not find container \"a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb\": container with ID starting with a5d617de90caf15b78c54eace0ed599c6ec5cee34a26956b2071760cb3a927bb not found: ID does not exist" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.266457 4989 scope.go:117] "RemoveContainer" containerID="e5b348713785110cf8d1a2640e5c957ea70b13112bd34028c312f91a34d04d84" Oct 06 11:28:05 crc kubenswrapper[4989]: I1006 11:28:05.964945 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" path="/var/lib/kubelet/pods/aba00a00-6135-4e37-b0a0-69fd11d45bbc/volumes" Oct 06 11:28:09 crc kubenswrapper[4989]: I1006 11:28:09.115644 4989 generic.go:334] "Generic (PLEG): container finished" podID="efcb7c88-0b73-49a4-8b6d-67f67b2e25da" containerID="559f83e9826e64082e2b36bac21f6d13d22fad55dc0cd6585609126a645c8a8e" exitCode=0 Oct 06 11:28:09 crc kubenswrapper[4989]: I1006 11:28:09.115730 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" event={"ID":"efcb7c88-0b73-49a4-8b6d-67f67b2e25da","Type":"ContainerDied","Data":"559f83e9826e64082e2b36bac21f6d13d22fad55dc0cd6585609126a645c8a8e"} Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.174434 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" event={"ID":"efcb7c88-0b73-49a4-8b6d-67f67b2e25da","Type":"ContainerDied","Data":"c6229e3e22b0398d5324171ede81695a6457720d432e61d288e784a07adabf0c"} Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.175145 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6229e3e22b0398d5324171ede81695a6457720d432e61d288e784a07adabf0c" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.188966 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.235307 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rfz2j"] Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.245472 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rfz2j"] Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.309709 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnlzb\" (UniqueName: \"kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb\") pod \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.309885 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host\") pod \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\" (UID: \"efcb7c88-0b73-49a4-8b6d-67f67b2e25da\") " Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.310108 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host" (OuterVolumeSpecName: "host") pod "efcb7c88-0b73-49a4-8b6d-67f67b2e25da" (UID: "efcb7c88-0b73-49a4-8b6d-67f67b2e25da"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.310499 4989 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.322581 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb" (OuterVolumeSpecName: "kube-api-access-xnlzb") pod "efcb7c88-0b73-49a4-8b6d-67f67b2e25da" (UID: "efcb7c88-0b73-49a4-8b6d-67f67b2e25da"). InnerVolumeSpecName "kube-api-access-xnlzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.415028 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnlzb\" (UniqueName: \"kubernetes.io/projected/efcb7c88-0b73-49a4-8b6d-67f67b2e25da-kube-api-access-xnlzb\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:11 crc kubenswrapper[4989]: I1006 11:28:11.953381 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efcb7c88-0b73-49a4-8b6d-67f67b2e25da" path="/var/lib/kubelet/pods/efcb7c88-0b73-49a4-8b6d-67f67b2e25da/volumes" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.187134 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rfz2j" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.431139 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-nq6l5"] Oct 06 11:28:12 crc kubenswrapper[4989]: E1006 11:28:12.431789 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="extract-utilities" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.431821 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="extract-utilities" Oct 06 11:28:12 crc kubenswrapper[4989]: E1006 11:28:12.431846 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcb7c88-0b73-49a4-8b6d-67f67b2e25da" containerName="container-00" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.431855 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcb7c88-0b73-49a4-8b6d-67f67b2e25da" containerName="container-00" Oct 06 11:28:12 crc kubenswrapper[4989]: E1006 11:28:12.431882 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="extract-content" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.431893 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="extract-content" Oct 06 11:28:12 crc kubenswrapper[4989]: E1006 11:28:12.431960 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="registry-server" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.431971 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="registry-server" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.432246 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcb7c88-0b73-49a4-8b6d-67f67b2e25da" containerName="container-00" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.432284 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="aba00a00-6135-4e37-b0a0-69fd11d45bbc" containerName="registry-server" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.433295 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.435928 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qvnb8"/"default-dockercfg-2n8kk" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.547730 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.548566 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4gvb\" (UniqueName: \"kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.651424 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4gvb\" (UniqueName: \"kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.651618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.651789 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.676942 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4gvb\" (UniqueName: \"kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb\") pod \"crc-debug-nq6l5\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: I1006 11:28:12.757252 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:12 crc kubenswrapper[4989]: W1006 11:28:12.813870 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3168b2ee_a5d2_40bb_b5de_09f3376a9058.slice/crio-1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b WatchSource:0}: Error finding container 1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b: Status 404 returned error can't find the container with id 1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b Oct 06 11:28:13 crc kubenswrapper[4989]: I1006 11:28:13.204286 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" event={"ID":"3168b2ee-a5d2-40bb-b5de-09f3376a9058","Type":"ContainerStarted","Data":"9233e16fe3761351e7e392ec735a882dc304fdb336fd058083723b6ea8e10d82"} Oct 06 11:28:13 crc kubenswrapper[4989]: I1006 11:28:13.204957 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" event={"ID":"3168b2ee-a5d2-40bb-b5de-09f3376a9058","Type":"ContainerStarted","Data":"1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b"} Oct 06 11:28:13 crc kubenswrapper[4989]: I1006 11:28:13.244102 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" podStartSLOduration=1.244067281 podStartE2EDuration="1.244067281s" podCreationTimestamp="2025-10-06 11:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:28:13.224037864 +0000 UTC m=+10144.014063504" watchObservedRunningTime="2025-10-06 11:28:13.244067281 +0000 UTC m=+10144.034092931" Oct 06 11:28:14 crc kubenswrapper[4989]: I1006 11:28:14.214754 4989 generic.go:334] "Generic (PLEG): container finished" podID="3168b2ee-a5d2-40bb-b5de-09f3376a9058" containerID="9233e16fe3761351e7e392ec735a882dc304fdb336fd058083723b6ea8e10d82" exitCode=0 Oct 06 11:28:14 crc kubenswrapper[4989]: I1006 11:28:14.214826 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" event={"ID":"3168b2ee-a5d2-40bb-b5de-09f3376a9058","Type":"ContainerDied","Data":"9233e16fe3761351e7e392ec735a882dc304fdb336fd058083723b6ea8e10d82"} Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.333242 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.418031 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4gvb\" (UniqueName: \"kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb\") pod \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.418200 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host\") pod \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\" (UID: \"3168b2ee-a5d2-40bb-b5de-09f3376a9058\") " Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.418319 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host" (OuterVolumeSpecName: "host") pod "3168b2ee-a5d2-40bb-b5de-09f3376a9058" (UID: "3168b2ee-a5d2-40bb-b5de-09f3376a9058"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.419419 4989 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3168b2ee-a5d2-40bb-b5de-09f3376a9058-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.427086 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb" (OuterVolumeSpecName: "kube-api-access-p4gvb") pod "3168b2ee-a5d2-40bb-b5de-09f3376a9058" (UID: "3168b2ee-a5d2-40bb-b5de-09f3376a9058"). InnerVolumeSpecName "kube-api-access-p4gvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:28:15 crc kubenswrapper[4989]: I1006 11:28:15.521800 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4gvb\" (UniqueName: \"kubernetes.io/projected/3168b2ee-a5d2-40bb-b5de-09f3376a9058-kube-api-access-p4gvb\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:16 crc kubenswrapper[4989]: I1006 11:28:16.252863 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" event={"ID":"3168b2ee-a5d2-40bb-b5de-09f3376a9058","Type":"ContainerDied","Data":"1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b"} Oct 06 11:28:16 crc kubenswrapper[4989]: I1006 11:28:16.252917 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c194cf3bf474c4eddc9d4e56892553654a9396d0cd1c299197b02e8675fdb5b" Oct 06 11:28:16 crc kubenswrapper[4989]: I1006 11:28:16.253005 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-nq6l5" Oct 06 11:28:25 crc kubenswrapper[4989]: I1006 11:28:25.723745 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-nq6l5"] Oct 06 11:28:25 crc kubenswrapper[4989]: I1006 11:28:25.746862 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-nq6l5"] Oct 06 11:28:25 crc kubenswrapper[4989]: I1006 11:28:25.958889 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3168b2ee-a5d2-40bb-b5de-09f3376a9058" path="/var/lib/kubelet/pods/3168b2ee-a5d2-40bb-b5de-09f3376a9058/volumes" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.005583 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rqxl2"] Oct 06 11:28:27 crc kubenswrapper[4989]: E1006 11:28:27.007007 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3168b2ee-a5d2-40bb-b5de-09f3376a9058" containerName="container-00" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.007036 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="3168b2ee-a5d2-40bb-b5de-09f3376a9058" containerName="container-00" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.007517 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="3168b2ee-a5d2-40bb-b5de-09f3376a9058" containerName="container-00" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.009028 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.012496 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qvnb8"/"default-dockercfg-2n8kk" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.124236 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f58dr\" (UniqueName: \"kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.124532 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.226618 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.226814 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f58dr\" (UniqueName: \"kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.226832 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.262497 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f58dr\" (UniqueName: \"kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr\") pod \"crc-debug-rqxl2\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.336143 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:27 crc kubenswrapper[4989]: I1006 11:28:27.452433 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" event={"ID":"87bc3055-36ef-4ca1-b46b-744fbb420a75","Type":"ContainerStarted","Data":"6a061dbf370e59b8070ca0720ac26c2698fb59733f2a9093c50e4c430ec6d94e"} Oct 06 11:28:28 crc kubenswrapper[4989]: I1006 11:28:28.471068 4989 generic.go:334] "Generic (PLEG): container finished" podID="87bc3055-36ef-4ca1-b46b-744fbb420a75" containerID="55c0dbd967a049dd8f060872156427606166c0c0bb052c8c1389a0a1b5b04de3" exitCode=0 Oct 06 11:28:28 crc kubenswrapper[4989]: I1006 11:28:28.471165 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" event={"ID":"87bc3055-36ef-4ca1-b46b-744fbb420a75","Type":"ContainerDied","Data":"55c0dbd967a049dd8f060872156427606166c0c0bb052c8c1389a0a1b5b04de3"} Oct 06 11:28:28 crc kubenswrapper[4989]: I1006 11:28:28.536486 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rqxl2"] Oct 06 11:28:28 crc kubenswrapper[4989]: I1006 11:28:28.546948 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qvnb8/crc-debug-rqxl2"] Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.611547 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.699294 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host\") pod \"87bc3055-36ef-4ca1-b46b-744fbb420a75\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.699424 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f58dr\" (UniqueName: \"kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr\") pod \"87bc3055-36ef-4ca1-b46b-744fbb420a75\" (UID: \"87bc3055-36ef-4ca1-b46b-744fbb420a75\") " Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.699409 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host" (OuterVolumeSpecName: "host") pod "87bc3055-36ef-4ca1-b46b-744fbb420a75" (UID: "87bc3055-36ef-4ca1-b46b-744fbb420a75"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.700358 4989 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87bc3055-36ef-4ca1-b46b-744fbb420a75-host\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.708080 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr" (OuterVolumeSpecName: "kube-api-access-f58dr") pod "87bc3055-36ef-4ca1-b46b-744fbb420a75" (UID: "87bc3055-36ef-4ca1-b46b-744fbb420a75"). InnerVolumeSpecName "kube-api-access-f58dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.803403 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f58dr\" (UniqueName: \"kubernetes.io/projected/87bc3055-36ef-4ca1-b46b-744fbb420a75-kube-api-access-f58dr\") on node \"crc\" DevicePath \"\"" Oct 06 11:28:29 crc kubenswrapper[4989]: I1006 11:28:29.960324 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87bc3055-36ef-4ca1-b46b-744fbb420a75" path="/var/lib/kubelet/pods/87bc3055-36ef-4ca1-b46b-744fbb420a75/volumes" Oct 06 11:28:30 crc kubenswrapper[4989]: I1006 11:28:30.506684 4989 scope.go:117] "RemoveContainer" containerID="55c0dbd967a049dd8f060872156427606166c0c0bb052c8c1389a0a1b5b04de3" Oct 06 11:28:30 crc kubenswrapper[4989]: I1006 11:28:30.507353 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/crc-debug-rqxl2" Oct 06 11:28:49 crc kubenswrapper[4989]: I1006 11:28:49.821151 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-4skd5_9ed264b3-7650-41ed-bb45-c5e0265ebcaf/kube-rbac-proxy/0.log" Oct 06 11:28:49 crc kubenswrapper[4989]: I1006 11:28:49.944090 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-4skd5_9ed264b3-7650-41ed-bb45-c5e0265ebcaf/manager/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.083678 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/util/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.306156 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/util/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.320482 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/pull/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.320610 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/pull/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.511183 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/util/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.516925 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/extract/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.521099 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cc5c9d72985612613b60d673fcd9a5546868deebd03ee102aed9b84bcf6shsd_ba2e51bd-d38a-41e4-91eb-853d18b5a459/pull/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.714591 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-6r4p2_9e971444-f7c2-4443-ad67-8344cc2ece01/kube-rbac-proxy/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.763231 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-8h5sx_c0f513bb-5736-42bd-a6cb-cf5314b940ee/kube-rbac-proxy/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.848698 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-6r4p2_9e971444-f7c2-4443-ad67-8344cc2ece01/manager/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.924173 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-8h5sx_c0f513bb-5736-42bd-a6cb-cf5314b940ee/manager/0.log" Oct 06 11:28:50 crc kubenswrapper[4989]: I1006 11:28:50.981171 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-25cxf_cce21119-33b5-4f7c-834c-0e8f16ccd825/kube-rbac-proxy/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.227931 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-phj9c_30e9d8ad-e772-44e4-8234-a0ef8a3483e1/kube-rbac-proxy/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.281765 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-25cxf_cce21119-33b5-4f7c-834c-0e8f16ccd825/manager/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.302384 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-phj9c_30e9d8ad-e772-44e4-8234-a0ef8a3483e1/manager/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.396182 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-dqjpg_b1667c4d-4d3a-41b1-a02a-890a2b0f26ea/kube-rbac-proxy/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.495248 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-dqjpg_b1667c4d-4d3a-41b1-a02a-890a2b0f26ea/manager/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.587522 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cdzbs_05b084da-1f15-4dba-8d4e-b540d4447318/kube-rbac-proxy/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.743421 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-mrz27_b89d614b-bb9f-4ba5-bb16-d9fe9731279f/kube-rbac-proxy/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.849683 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-mrz27_b89d614b-bb9f-4ba5-bb16-d9fe9731279f/manager/0.log" Oct 06 11:28:51 crc kubenswrapper[4989]: I1006 11:28:51.952497 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-cdzbs_05b084da-1f15-4dba-8d4e-b540d4447318/manager/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.020079 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-79ksl_45344ba0-bd44-4a54-9628-460586eae8df/kube-rbac-proxy/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.168609 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gmfps_9477ed03-7f88-4abf-bbcd-65a2888aff71/kube-rbac-proxy/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.202600 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-79ksl_45344ba0-bd44-4a54-9628-460586eae8df/manager/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.289703 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-gmfps_9477ed03-7f88-4abf-bbcd-65a2888aff71/manager/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.404286 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-m78db_9acc4601-f863-47b9-8a11-d91615ab1af5/kube-rbac-proxy/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.505148 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-m78db_9acc4601-f863-47b9-8a11-d91615ab1af5/manager/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.599703 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fff4f_290501cd-b2bb-4520-b8b0-0197017e61c6/kube-rbac-proxy/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.661590 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fff4f_290501cd-b2bb-4520-b8b0-0197017e61c6/manager/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.753355 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-997ct_70bea185-7eb6-4fc4-8016-c32e11078df1/kube-rbac-proxy/0.log" Oct 06 11:28:52 crc kubenswrapper[4989]: I1006 11:28:52.999613 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-jgjms_9d82b2b6-9374-4cbc-a72c-50aa000a0d52/kube-rbac-proxy/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.151333 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-jgjms_9d82b2b6-9374-4cbc-a72c-50aa000a0d52/manager/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.154659 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-997ct_70bea185-7eb6-4fc4-8016-c32e11078df1/manager/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.291079 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz_3af9ba41-3325-4e75-a392-306b9999606b/kube-rbac-proxy/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.393625 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665cc9ffz_3af9ba41-3325-4e75-a392-306b9999606b/manager/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.410100 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-669d7f654d-rvb7c_7fba2f19-add5-4c43-a142-82dd5c0ba564/kube-rbac-proxy/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.699086 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6497dff45c-mpcvt_2b3e23ba-5e26-49ef-acaf-0520111b142a/kube-rbac-proxy/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.823384 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jjkjt_3c87cc47-079e-44f3-bac3-a66ce98b1692/registry-server/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.861781 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6497dff45c-mpcvt_2b3e23ba-5e26-49ef-acaf-0520111b142a/operator/0.log" Oct 06 11:28:53 crc kubenswrapper[4989]: I1006 11:28:53.919488 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-lzjq2_21ae0953-021f-40d1-aab0-2b693f29212a/kube-rbac-proxy/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.202995 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-lzjq2_21ae0953-021f-40d1-aab0-2b693f29212a/manager/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.227076 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-l62dg_97f22726-47c7-47db-b9a2-9efbddfaa7ba/kube-rbac-proxy/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.232408 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-l62dg_97f22726-47c7-47db-b9a2-9efbddfaa7ba/manager/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.565153 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-kzsql_9168e44a-597f-4623-9520-b9fe4c46db99/operator/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.621673 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-kgbzk_522e1bb1-d22f-4256-b0f0-ae27447bd00a/kube-rbac-proxy/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.720284 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-kgbzk_522e1bb1-d22f-4256-b0f0-ae27447bd00a/manager/0.log" Oct 06 11:28:54 crc kubenswrapper[4989]: I1006 11:28:54.762230 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-nm6z6_a3213843-170f-4e4b-b35b-87f424dd2abb/kube-rbac-proxy/0.log" Oct 06 11:28:55 crc kubenswrapper[4989]: I1006 11:28:55.038167 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-b5kt7_d8ec36ee-cada-48de-b02d-69c1c21ae7b2/manager/0.log" Oct 06 11:28:55 crc kubenswrapper[4989]: I1006 11:28:55.101518 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-b5kt7_d8ec36ee-cada-48de-b02d-69c1c21ae7b2/kube-rbac-proxy/0.log" Oct 06 11:28:55 crc kubenswrapper[4989]: I1006 11:28:55.241933 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-672zz_428077cb-5a71-4f65-a37e-2384c4c0af19/kube-rbac-proxy/0.log" Oct 06 11:28:55 crc kubenswrapper[4989]: I1006 11:28:55.297871 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-nm6z6_a3213843-170f-4e4b-b35b-87f424dd2abb/manager/0.log" Oct 06 11:28:55 crc kubenswrapper[4989]: I1006 11:28:55.369418 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-672zz_428077cb-5a71-4f65-a37e-2384c4c0af19/manager/0.log" Oct 06 11:28:56 crc kubenswrapper[4989]: I1006 11:28:56.207412 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-669d7f654d-rvb7c_7fba2f19-add5-4c43-a142-82dd5c0ba564/manager/0.log" Oct 06 11:29:17 crc kubenswrapper[4989]: I1006 11:29:17.078793 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hr5wj_7dcad5af-901c-431f-b7b9-a98b4e667ee8/control-plane-machine-set-operator/0.log" Oct 06 11:29:17 crc kubenswrapper[4989]: I1006 11:29:17.262528 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-m7lpj_9ec320c3-776d-4197-9259-7feee562229a/kube-rbac-proxy/0.log" Oct 06 11:29:17 crc kubenswrapper[4989]: I1006 11:29:17.390606 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-m7lpj_9ec320c3-776d-4197-9259-7feee562229a/machine-api-operator/0.log" Oct 06 11:29:32 crc kubenswrapper[4989]: I1006 11:29:32.792549 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-pzq52_f0a0cf7b-ef3f-4b08-a574-9f81f13d227b/cert-manager-controller/0.log" Oct 06 11:29:32 crc kubenswrapper[4989]: I1006 11:29:32.911205 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-5d6bh_6be97c71-f240-44ab-bba5-f52434740520/cert-manager-cainjector/0.log" Oct 06 11:29:32 crc kubenswrapper[4989]: I1006 11:29:32.994940 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-9djb6_48613a5d-3bd1-4731-ab29-efe554110a70/cert-manager-webhook/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.474463 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-fglz7_2ca0d980-4342-41a1-a47a-d8c55d87254c/nmstate-console-plugin/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.651628 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-j7w6r_e62c6594-51c2-42d8-9b9e-d8ed53a60be4/nmstate-handler/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.686112 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-85qpx_734fb7db-406a-47af-ae7b-5e346452ce57/kube-rbac-proxy/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.756803 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-85qpx_734fb7db-406a-47af-ae7b-5e346452ce57/nmstate-metrics/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.910047 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-2btck_4f4c242d-e5b1-4091-9960-9c7e7141f04f/nmstate-operator/0.log" Oct 06 11:29:48 crc kubenswrapper[4989]: I1006 11:29:48.966319 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-sx7g9_fc10f366-5738-466e-96ac-352e579427aa/nmstate-webhook/0.log" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.179031 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh"] Oct 06 11:30:00 crc kubenswrapper[4989]: E1006 11:30:00.180071 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bc3055-36ef-4ca1-b46b-744fbb420a75" containerName="container-00" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.180088 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bc3055-36ef-4ca1-b46b-744fbb420a75" containerName="container-00" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.180456 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="87bc3055-36ef-4ca1-b46b-744fbb420a75" containerName="container-00" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.181713 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.184623 4989 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.193257 4989 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.193885 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh"] Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.325720 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rczps\" (UniqueName: \"kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.325835 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.325885 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.428132 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczps\" (UniqueName: \"kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.428211 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.428243 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.429377 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.785583 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.793580 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczps\" (UniqueName: \"kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps\") pod \"collect-profiles-29329170-2rjrh\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:00 crc kubenswrapper[4989]: I1006 11:30:00.815014 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:01 crc kubenswrapper[4989]: I1006 11:30:01.292753 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh"] Oct 06 11:30:01 crc kubenswrapper[4989]: I1006 11:30:01.685022 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" event={"ID":"32c7b1c3-5d35-4645-8f12-05219e71e9f0","Type":"ContainerStarted","Data":"a1093ded6cdaac2449b89c41f6d1edb6130b5bf524ce70786da415c0de9053e2"} Oct 06 11:30:01 crc kubenswrapper[4989]: I1006 11:30:01.685369 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" event={"ID":"32c7b1c3-5d35-4645-8f12-05219e71e9f0","Type":"ContainerStarted","Data":"e14baf57924718274e69f153d50055223edecb1a1d0490eac3f78c07656fafa6"} Oct 06 11:30:01 crc kubenswrapper[4989]: I1006 11:30:01.707998 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" podStartSLOduration=1.707976763 podStartE2EDuration="1.707976763s" podCreationTimestamp="2025-10-06 11:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-06 11:30:01.703561036 +0000 UTC m=+10252.493586616" watchObservedRunningTime="2025-10-06 11:30:01.707976763 +0000 UTC m=+10252.498002353" Oct 06 11:30:02 crc kubenswrapper[4989]: I1006 11:30:02.702366 4989 generic.go:334] "Generic (PLEG): container finished" podID="32c7b1c3-5d35-4645-8f12-05219e71e9f0" containerID="a1093ded6cdaac2449b89c41f6d1edb6130b5bf524ce70786da415c0de9053e2" exitCode=0 Oct 06 11:30:02 crc kubenswrapper[4989]: I1006 11:30:02.702445 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" event={"ID":"32c7b1c3-5d35-4645-8f12-05219e71e9f0","Type":"ContainerDied","Data":"a1093ded6cdaac2449b89c41f6d1edb6130b5bf524ce70786da415c0de9053e2"} Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.156233 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.224391 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rczps\" (UniqueName: \"kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps\") pod \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.224676 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume\") pod \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.224851 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume\") pod \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\" (UID: \"32c7b1c3-5d35-4645-8f12-05219e71e9f0\") " Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.225430 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "32c7b1c3-5d35-4645-8f12-05219e71e9f0" (UID: "32c7b1c3-5d35-4645-8f12-05219e71e9f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.233813 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "32c7b1c3-5d35-4645-8f12-05219e71e9f0" (UID: "32c7b1c3-5d35-4645-8f12-05219e71e9f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.233891 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps" (OuterVolumeSpecName: "kube-api-access-rczps") pod "32c7b1c3-5d35-4645-8f12-05219e71e9f0" (UID: "32c7b1c3-5d35-4645-8f12-05219e71e9f0"). InnerVolumeSpecName "kube-api-access-rczps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.327448 4989 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/32c7b1c3-5d35-4645-8f12-05219e71e9f0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.327482 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rczps\" (UniqueName: \"kubernetes.io/projected/32c7b1c3-5d35-4645-8f12-05219e71e9f0-kube-api-access-rczps\") on node \"crc\" DevicePath \"\"" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.327492 4989 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/32c7b1c3-5d35-4645-8f12-05219e71e9f0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.354664 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw"] Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.364813 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29329125-kbtmw"] Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.732627 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" event={"ID":"32c7b1c3-5d35-4645-8f12-05219e71e9f0","Type":"ContainerDied","Data":"e14baf57924718274e69f153d50055223edecb1a1d0490eac3f78c07656fafa6"} Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.732785 4989 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e14baf57924718274e69f153d50055223edecb1a1d0490eac3f78c07656fafa6" Oct 06 11:30:04 crc kubenswrapper[4989]: I1006 11:30:04.732694 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29329170-2rjrh" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.360306 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cdb7w_cfd98297-3ba6-455c-b039-8fb182d89aa0/kube-rbac-proxy/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.597939 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-frr-files/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.799533 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-reloader/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.848904 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-frr-files/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.858300 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cdb7w_cfd98297-3ba6-455c-b039-8fb182d89aa0/controller/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.901055 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-metrics/0.log" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.951280 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc87c85-b9ee-48b5-a357-015a136c798e" path="/var/lib/kubelet/pods/3bc87c85-b9ee-48b5-a357-015a136c798e/volumes" Oct 06 11:30:05 crc kubenswrapper[4989]: I1006 11:30:05.995270 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-reloader/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.182839 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-reloader/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.187698 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-metrics/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.204644 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-frr-files/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.226517 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-metrics/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.405233 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-reloader/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.413034 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-frr-files/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.429491 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/cp-metrics/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.436298 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/controller/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.592121 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/frr-metrics/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.620703 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/kube-rbac-proxy-frr/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.633646 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/kube-rbac-proxy/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.831428 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/reloader/0.log" Oct 06 11:30:06 crc kubenswrapper[4989]: I1006 11:30:06.853867 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-bs986_a02ef7c3-577f-4284-88c6-ca16b3b335b8/frr-k8s-webhook-server/0.log" Oct 06 11:30:07 crc kubenswrapper[4989]: I1006 11:30:07.094695 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-745c6df699-vqlnn_342ae3db-e559-4882-bc4b-6999d8cb1a13/manager/0.log" Oct 06 11:30:07 crc kubenswrapper[4989]: I1006 11:30:07.273894 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-59f68d59f5-cv5nf_9ff68b6e-9a02-4f31-8111-564bae539f3a/webhook-server/0.log" Oct 06 11:30:07 crc kubenswrapper[4989]: I1006 11:30:07.488053 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sqqsn_c021ef53-6c43-47fe-8429-f374626d4611/kube-rbac-proxy/0.log" Oct 06 11:30:08 crc kubenswrapper[4989]: I1006 11:30:08.406835 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sqqsn_c021ef53-6c43-47fe-8429-f374626d4611/speaker/0.log" Oct 06 11:30:10 crc kubenswrapper[4989]: I1006 11:30:10.083148 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gxdgx_38c9054e-238b-4eef-8f71-ec00117efac9/frr/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.559446 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/util/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.690968 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/util/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.712317 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/pull/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.738221 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/pull/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.894625 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/util/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.945766 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/extract/0.log" Oct 06 11:30:23 crc kubenswrapper[4989]: I1006 11:30:23.964173 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69mw57c_180f1859-bedd-4e03-81f5-553e5bf46f2b/pull/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.106838 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/util/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.653409 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/util/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.678315 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/pull/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.750234 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/pull/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.837353 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/util/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.888687 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/pull/0.log" Oct 06 11:30:24 crc kubenswrapper[4989]: I1006 11:30:24.903804 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2kp8x9_4f4d3ca6-b70d-465c-9a76-a98e046542c4/extract/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.046557 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/util/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.253730 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/util/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.255794 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/pull/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.284246 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/pull/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.435977 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/util/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.438788 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/extract/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.511533 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dwvnlj_ae45c68d-94f2-4ce1-a089-eb71c94e3543/pull/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.655791 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-utilities/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.868203 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-utilities/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.868425 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-content/0.log" Oct 06 11:30:25 crc kubenswrapper[4989]: I1006 11:30:25.910909 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-content/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.102950 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-content/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.123106 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/extract-utilities/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.355182 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-utilities/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.907476 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-utilities/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.915343 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-content/0.log" Oct 06 11:30:26 crc kubenswrapper[4989]: I1006 11:30:26.970344 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-content/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.096593 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-utilities/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.170993 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pwzh2_0b11d153-4f54-4a5e-9a21-69527194a1e4/registry-server/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.179201 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/extract-content/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.385002 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/util/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.559397 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/pull/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.567916 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/util/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.611904 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/pull/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.833645 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/extract/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.863109 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/pull/0.log" Oct 06 11:30:27 crc kubenswrapper[4989]: I1006 11:30:27.963866 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cnb52g_93fa8693-a1bc-4438-8a3c-318c6374f1b4/util/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.141589 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xkxs2_fd477c82-fc0b-4d4a-8de4-037f8f30e0e3/marketplace-operator/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.160647 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-utilities/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.431711 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-content/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.442700 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-utilities/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.468859 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-content/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.650307 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-content/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.682284 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/extract-utilities/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.866571 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-utilities/0.log" Oct 06 11:30:28 crc kubenswrapper[4989]: I1006 11:30:28.969182 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-cjkm6_aaae90ae-1c87-4233-85aa-98d3dbbc7790/registry-server/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.087636 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-content/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.124408 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-utilities/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.126953 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vxfgz_50b0a5d8-8aa4-45e4-bf00-c6e717c3c1bf/registry-server/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.154073 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-content/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.271387 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-utilities/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.275147 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/extract-content/0.log" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.581131 4989 scope.go:117] "RemoveContainer" containerID="4e91ae88869d41b812769fbd27c3c30a418ac5e421058a59d886552807444638" Oct 06 11:30:29 crc kubenswrapper[4989]: I1006 11:30:29.587580 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vdqv9_53b2dca6-6028-4b82-80bc-5b3109054379/registry-server/0.log" Oct 06 11:30:33 crc kubenswrapper[4989]: I1006 11:30:33.938825 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:30:33 crc kubenswrapper[4989]: I1006 11:30:33.939860 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:30:43 crc kubenswrapper[4989]: I1006 11:30:43.698577 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-xmbvr_de27d14c-6b73-4043-a3ca-240a7e9c764f/prometheus-operator/0.log" Oct 06 11:30:43 crc kubenswrapper[4989]: I1006 11:30:43.792844 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7654b8595-7bkv2_c082c7ef-21ff-4e43-b97c-f7b900a4a304/prometheus-operator-admission-webhook/0.log" Oct 06 11:30:43 crc kubenswrapper[4989]: I1006 11:30:43.906148 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7654b8595-7g6c9_3bf294dc-d953-4dc7-afcf-939c59d64ffa/prometheus-operator-admission-webhook/0.log" Oct 06 11:30:44 crc kubenswrapper[4989]: I1006 11:30:44.037738 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-sqmxd_db80ab6a-4df2-47ad-9883-83dd1e730f01/operator/0.log" Oct 06 11:30:44 crc kubenswrapper[4989]: I1006 11:30:44.127138 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-dfztc_df7a3bee-462c-4a74-8449-1ea82af24bdd/perses-operator/0.log" Oct 06 11:30:59 crc kubenswrapper[4989]: E1006 11:30:59.511814 4989 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.23:39644->38.129.56.23:40321: write tcp 38.129.56.23:39644->38.129.56.23:40321: write: broken pipe Oct 06 11:31:03 crc kubenswrapper[4989]: I1006 11:31:03.935879 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:31:03 crc kubenswrapper[4989]: I1006 11:31:03.936346 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:31:33 crc kubenswrapper[4989]: I1006 11:31:33.936697 4989 patch_prober.go:28] interesting pod/machine-config-daemon-v6vsq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 06 11:31:33 crc kubenswrapper[4989]: I1006 11:31:33.938325 4989 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 06 11:31:33 crc kubenswrapper[4989]: I1006 11:31:33.949010 4989 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" Oct 06 11:31:33 crc kubenswrapper[4989]: I1006 11:31:33.949973 4989 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36"} pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 06 11:31:33 crc kubenswrapper[4989]: I1006 11:31:33.950050 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerName="machine-config-daemon" containerID="cri-o://5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" gracePeriod=600 Oct 06 11:31:34 crc kubenswrapper[4989]: E1006 11:31:34.597111 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:31:34 crc kubenswrapper[4989]: I1006 11:31:34.864330 4989 generic.go:334] "Generic (PLEG): container finished" podID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" exitCode=0 Oct 06 11:31:34 crc kubenswrapper[4989]: I1006 11:31:34.864384 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" event={"ID":"f78ef9f2-a706-4a6c-8a9f-b232f97e9211","Type":"ContainerDied","Data":"5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36"} Oct 06 11:31:34 crc kubenswrapper[4989]: I1006 11:31:34.864430 4989 scope.go:117] "RemoveContainer" containerID="5c98c4495932941c7939273edfa1d219d1fbdac293e53b1e91048fa9af920052" Oct 06 11:31:34 crc kubenswrapper[4989]: I1006 11:31:34.866150 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:31:34 crc kubenswrapper[4989]: E1006 11:31:34.866714 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:31:47 crc kubenswrapper[4989]: I1006 11:31:47.936023 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:31:47 crc kubenswrapper[4989]: E1006 11:31:47.936810 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:32:02 crc kubenswrapper[4989]: I1006 11:32:02.937009 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:32:02 crc kubenswrapper[4989]: E1006 11:32:02.938036 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:32:17 crc kubenswrapper[4989]: I1006 11:32:17.936171 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:32:17 crc kubenswrapper[4989]: E1006 11:32:17.937337 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:32:29 crc kubenswrapper[4989]: I1006 11:32:29.682714 4989 scope.go:117] "RemoveContainer" containerID="559f83e9826e64082e2b36bac21f6d13d22fad55dc0cd6585609126a645c8a8e" Oct 06 11:32:29 crc kubenswrapper[4989]: I1006 11:32:29.951624 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:32:29 crc kubenswrapper[4989]: E1006 11:32:29.952250 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:32:41 crc kubenswrapper[4989]: I1006 11:32:41.937250 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:32:41 crc kubenswrapper[4989]: E1006 11:32:41.938854 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:32:52 crc kubenswrapper[4989]: I1006 11:32:52.937987 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:32:52 crc kubenswrapper[4989]: E1006 11:32:52.940150 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:33:07 crc kubenswrapper[4989]: I1006 11:33:07.937256 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:33:07 crc kubenswrapper[4989]: E1006 11:33:07.938178 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:33:18 crc kubenswrapper[4989]: I1006 11:33:18.936178 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:33:18 crc kubenswrapper[4989]: E1006 11:33:18.937138 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:33:33 crc kubenswrapper[4989]: I1006 11:33:33.524166 4989 generic.go:334] "Generic (PLEG): container finished" podID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerID="391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4" exitCode=0 Oct 06 11:33:33 crc kubenswrapper[4989]: I1006 11:33:33.524246 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" event={"ID":"bd58a75a-cc5b-489b-ac94-57daa6f651ae","Type":"ContainerDied","Data":"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4"} Oct 06 11:33:33 crc kubenswrapper[4989]: I1006 11:33:33.526187 4989 scope.go:117] "RemoveContainer" containerID="391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4" Oct 06 11:33:33 crc kubenswrapper[4989]: I1006 11:33:33.937623 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:33:33 crc kubenswrapper[4989]: E1006 11:33:33.938829 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:33:34 crc kubenswrapper[4989]: I1006 11:33:34.405670 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvnb8_must-gather-rjtdk_bd58a75a-cc5b-489b-ac94-57daa6f651ae/gather/0.log" Oct 06 11:33:43 crc kubenswrapper[4989]: I1006 11:33:43.899707 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qvnb8/must-gather-rjtdk"] Oct 06 11:33:43 crc kubenswrapper[4989]: I1006 11:33:43.901081 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="copy" containerID="cri-o://73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace" gracePeriod=2 Oct 06 11:33:43 crc kubenswrapper[4989]: I1006 11:33:43.917804 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qvnb8/must-gather-rjtdk"] Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.378380 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvnb8_must-gather-rjtdk_bd58a75a-cc5b-489b-ac94-57daa6f651ae/copy/0.log" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.379928 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.415204 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxbfn\" (UniqueName: \"kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn\") pod \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.415448 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output\") pod \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\" (UID: \"bd58a75a-cc5b-489b-ac94-57daa6f651ae\") " Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.428487 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn" (OuterVolumeSpecName: "kube-api-access-dxbfn") pod "bd58a75a-cc5b-489b-ac94-57daa6f651ae" (UID: "bd58a75a-cc5b-489b-ac94-57daa6f651ae"). InnerVolumeSpecName "kube-api-access-dxbfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.518806 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxbfn\" (UniqueName: \"kubernetes.io/projected/bd58a75a-cc5b-489b-ac94-57daa6f651ae-kube-api-access-dxbfn\") on node \"crc\" DevicePath \"\"" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.664129 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "bd58a75a-cc5b-489b-ac94-57daa6f651ae" (UID: "bd58a75a-cc5b-489b-ac94-57daa6f651ae"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.712045 4989 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qvnb8_must-gather-rjtdk_bd58a75a-cc5b-489b-ac94-57daa6f651ae/copy/0.log" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.712457 4989 generic.go:334] "Generic (PLEG): container finished" podID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerID="73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace" exitCode=143 Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.712527 4989 scope.go:117] "RemoveContainer" containerID="73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.712776 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qvnb8/must-gather-rjtdk" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.738948 4989 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/bd58a75a-cc5b-489b-ac94-57daa6f651ae-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.758075 4989 scope.go:117] "RemoveContainer" containerID="391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.877876 4989 scope.go:117] "RemoveContainer" containerID="73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace" Oct 06 11:33:44 crc kubenswrapper[4989]: E1006 11:33:44.881708 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace\": container with ID starting with 73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace not found: ID does not exist" containerID="73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.881759 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace"} err="failed to get container status \"73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace\": rpc error: code = NotFound desc = could not find container \"73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace\": container with ID starting with 73dba71cc5e6541de9444a2db6c3bc4675ef1cd40790c8f304b4d54fe10b7ace not found: ID does not exist" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.881791 4989 scope.go:117] "RemoveContainer" containerID="391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4" Oct 06 11:33:44 crc kubenswrapper[4989]: E1006 11:33:44.882129 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4\": container with ID starting with 391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4 not found: ID does not exist" containerID="391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4" Oct 06 11:33:44 crc kubenswrapper[4989]: I1006 11:33:44.882150 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4"} err="failed to get container status \"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4\": rpc error: code = NotFound desc = could not find container \"391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4\": container with ID starting with 391dfe561fadd569f152dd8ad1d7deabf454d58d2c2b7e285662d1e2dcba5df4 not found: ID does not exist" Oct 06 11:33:45 crc kubenswrapper[4989]: I1006 11:33:45.950498 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" path="/var/lib/kubelet/pods/bd58a75a-cc5b-489b-ac94-57daa6f651ae/volumes" Oct 06 11:33:47 crc kubenswrapper[4989]: I1006 11:33:47.940578 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:33:47 crc kubenswrapper[4989]: E1006 11:33:47.941366 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:33:58 crc kubenswrapper[4989]: I1006 11:33:58.951355 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:33:58 crc kubenswrapper[4989]: E1006 11:33:58.954438 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:34:10 crc kubenswrapper[4989]: I1006 11:34:10.937630 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:34:10 crc kubenswrapper[4989]: E1006 11:34:10.939021 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:34:22 crc kubenswrapper[4989]: I1006 11:34:22.938898 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:34:22 crc kubenswrapper[4989]: E1006 11:34:22.939914 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:34:29 crc kubenswrapper[4989]: I1006 11:34:29.836694 4989 scope.go:117] "RemoveContainer" containerID="9233e16fe3761351e7e392ec735a882dc304fdb336fd058083723b6ea8e10d82" Oct 06 11:34:33 crc kubenswrapper[4989]: I1006 11:34:33.936300 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:34:33 crc kubenswrapper[4989]: E1006 11:34:33.937010 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.725290 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:34:42 crc kubenswrapper[4989]: E1006 11:34:42.731234 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c7b1c3-5d35-4645-8f12-05219e71e9f0" containerName="collect-profiles" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.731292 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c7b1c3-5d35-4645-8f12-05219e71e9f0" containerName="collect-profiles" Oct 06 11:34:42 crc kubenswrapper[4989]: E1006 11:34:42.731482 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="gather" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.731506 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="gather" Oct 06 11:34:42 crc kubenswrapper[4989]: E1006 11:34:42.731572 4989 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="copy" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.731599 4989 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="copy" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.736209 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c7b1c3-5d35-4645-8f12-05219e71e9f0" containerName="collect-profiles" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.736437 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="gather" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.736519 4989 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd58a75a-cc5b-489b-ac94-57daa6f651ae" containerName="copy" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.779403 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.779508 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.880579 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.880816 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.880865 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5mrq\" (UniqueName: \"kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.910374 4989 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.913181 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.922916 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.983051 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.983179 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.983261 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5mrq\" (UniqueName: \"kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.983315 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.984240 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.984292 4989 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npvqv\" (UniqueName: \"kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.984339 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:42 crc kubenswrapper[4989]: I1006 11:34:42.984729 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.006277 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5mrq\" (UniqueName: \"kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq\") pod \"community-operators-4rm8x\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.086128 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npvqv\" (UniqueName: \"kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.086480 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.086559 4989 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.087134 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.090822 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.109410 4989 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npvqv\" (UniqueName: \"kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv\") pod \"redhat-marketplace-phk7d\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.110345 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.232274 4989 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.647726 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:34:43 crc kubenswrapper[4989]: I1006 11:34:43.813509 4989 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:43 crc kubenswrapper[4989]: W1006 11:34:43.823481 4989 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod066b101a_f170_4506_a60f_f82236f05912.slice/crio-21c3e1843083deee1b8b5fb6d7422827ba1615e73f69b8044d7fb9ea5c4cbdaa WatchSource:0}: Error finding container 21c3e1843083deee1b8b5fb6d7422827ba1615e73f69b8044d7fb9ea5c4cbdaa: Status 404 returned error can't find the container with id 21c3e1843083deee1b8b5fb6d7422827ba1615e73f69b8044d7fb9ea5c4cbdaa Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.508785 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" containerID="6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d" exitCode=0 Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.508887 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerDied","Data":"6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d"} Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.509161 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerStarted","Data":"40c36c92bdeb5eab1e581ecb4a61b27ab8d113d7a636f5227105799cb3c11d59"} Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.511166 4989 generic.go:334] "Generic (PLEG): container finished" podID="066b101a-f170-4506-a60f-f82236f05912" containerID="c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8" exitCode=0 Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.511203 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerDied","Data":"c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8"} Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.511226 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerStarted","Data":"21c3e1843083deee1b8b5fb6d7422827ba1615e73f69b8044d7fb9ea5c4cbdaa"} Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.512545 4989 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 06 11:34:44 crc kubenswrapper[4989]: I1006 11:34:44.937455 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:34:44 crc kubenswrapper[4989]: E1006 11:34:44.938465 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:34:46 crc kubenswrapper[4989]: I1006 11:34:46.537439 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerStarted","Data":"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52"} Oct 06 11:34:46 crc kubenswrapper[4989]: I1006 11:34:46.540829 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerStarted","Data":"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d"} Oct 06 11:34:47 crc kubenswrapper[4989]: I1006 11:34:47.556274 4989 generic.go:334] "Generic (PLEG): container finished" podID="066b101a-f170-4506-a60f-f82236f05912" containerID="cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d" exitCode=0 Oct 06 11:34:47 crc kubenswrapper[4989]: I1006 11:34:47.556538 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerDied","Data":"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d"} Oct 06 11:34:48 crc kubenswrapper[4989]: I1006 11:34:48.568505 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" containerID="56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52" exitCode=0 Oct 06 11:34:48 crc kubenswrapper[4989]: I1006 11:34:48.569163 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerDied","Data":"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52"} Oct 06 11:34:48 crc kubenswrapper[4989]: I1006 11:34:48.575700 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerStarted","Data":"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f"} Oct 06 11:34:49 crc kubenswrapper[4989]: I1006 11:34:49.598367 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerStarted","Data":"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1"} Oct 06 11:34:49 crc kubenswrapper[4989]: I1006 11:34:49.628483 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-phk7d" podStartSLOduration=4.146943501 podStartE2EDuration="7.628454957s" podCreationTimestamp="2025-10-06 11:34:42 +0000 UTC" firstStartedPulling="2025-10-06 11:34:44.512995298 +0000 UTC m=+10535.303020888" lastFinishedPulling="2025-10-06 11:34:47.994506724 +0000 UTC m=+10538.784532344" observedRunningTime="2025-10-06 11:34:48.614811531 +0000 UTC m=+10539.404837141" watchObservedRunningTime="2025-10-06 11:34:49.628454957 +0000 UTC m=+10540.418480547" Oct 06 11:34:49 crc kubenswrapper[4989]: I1006 11:34:49.640099 4989 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4rm8x" podStartSLOduration=3.097459425 podStartE2EDuration="7.640072842s" podCreationTimestamp="2025-10-06 11:34:42 +0000 UTC" firstStartedPulling="2025-10-06 11:34:44.512308889 +0000 UTC m=+10535.302334469" lastFinishedPulling="2025-10-06 11:34:49.054922266 +0000 UTC m=+10539.844947886" observedRunningTime="2025-10-06 11:34:49.625702728 +0000 UTC m=+10540.415728348" watchObservedRunningTime="2025-10-06 11:34:49.640072842 +0000 UTC m=+10540.430098432" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.111572 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.112460 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.175555 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.233544 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.233620 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.313446 4989 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:53 crc kubenswrapper[4989]: I1006 11:34:53.726775 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:54 crc kubenswrapper[4989]: I1006 11:34:54.494072 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:55 crc kubenswrapper[4989]: I1006 11:34:55.675746 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-phk7d" podUID="066b101a-f170-4506-a60f-f82236f05912" containerName="registry-server" containerID="cri-o://751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f" gracePeriod=2 Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.209064 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.323046 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npvqv\" (UniqueName: \"kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv\") pod \"066b101a-f170-4506-a60f-f82236f05912\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.323226 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content\") pod \"066b101a-f170-4506-a60f-f82236f05912\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.323276 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities\") pod \"066b101a-f170-4506-a60f-f82236f05912\" (UID: \"066b101a-f170-4506-a60f-f82236f05912\") " Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.324413 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities" (OuterVolumeSpecName: "utilities") pod "066b101a-f170-4506-a60f-f82236f05912" (UID: "066b101a-f170-4506-a60f-f82236f05912"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.336891 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv" (OuterVolumeSpecName: "kube-api-access-npvqv") pod "066b101a-f170-4506-a60f-f82236f05912" (UID: "066b101a-f170-4506-a60f-f82236f05912"). InnerVolumeSpecName "kube-api-access-npvqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.342176 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "066b101a-f170-4506-a60f-f82236f05912" (UID: "066b101a-f170-4506-a60f-f82236f05912"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.425333 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.425581 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/066b101a-f170-4506-a60f-f82236f05912-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.425696 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npvqv\" (UniqueName: \"kubernetes.io/projected/066b101a-f170-4506-a60f-f82236f05912-kube-api-access-npvqv\") on node \"crc\" DevicePath \"\"" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.688911 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-phk7d" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.691293 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerDied","Data":"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f"} Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.692009 4989 scope.go:117] "RemoveContainer" containerID="751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.688906 4989 generic.go:334] "Generic (PLEG): container finished" podID="066b101a-f170-4506-a60f-f82236f05912" containerID="751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f" exitCode=0 Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.692563 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-phk7d" event={"ID":"066b101a-f170-4506-a60f-f82236f05912","Type":"ContainerDied","Data":"21c3e1843083deee1b8b5fb6d7422827ba1615e73f69b8044d7fb9ea5c4cbdaa"} Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.734403 4989 scope.go:117] "RemoveContainer" containerID="cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.750995 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.760813 4989 scope.go:117] "RemoveContainer" containerID="c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.766332 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-phk7d"] Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.828478 4989 scope.go:117] "RemoveContainer" containerID="751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f" Oct 06 11:34:56 crc kubenswrapper[4989]: E1006 11:34:56.829152 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f\": container with ID starting with 751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f not found: ID does not exist" containerID="751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.829243 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f"} err="failed to get container status \"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f\": rpc error: code = NotFound desc = could not find container \"751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f\": container with ID starting with 751c89d8fa465a64e7d4feb00b145a64019262cf5c819699ba7d6641608ed13f not found: ID does not exist" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.829304 4989 scope.go:117] "RemoveContainer" containerID="cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d" Oct 06 11:34:56 crc kubenswrapper[4989]: E1006 11:34:56.830285 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d\": container with ID starting with cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d not found: ID does not exist" containerID="cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.830318 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d"} err="failed to get container status \"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d\": rpc error: code = NotFound desc = could not find container \"cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d\": container with ID starting with cb9cb882fbb7270fb2e41f3717ad6abd3016997603bb8957b2acc5063f6a283d not found: ID does not exist" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.830337 4989 scope.go:117] "RemoveContainer" containerID="c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8" Oct 06 11:34:56 crc kubenswrapper[4989]: E1006 11:34:56.830709 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8\": container with ID starting with c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8 not found: ID does not exist" containerID="c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8" Oct 06 11:34:56 crc kubenswrapper[4989]: I1006 11:34:56.830758 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8"} err="failed to get container status \"c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8\": rpc error: code = NotFound desc = could not find container \"c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8\": container with ID starting with c778d827e3e8c79ae6693eba208423727b02c8f55733d2fd3f03fed3891e65a8 not found: ID does not exist" Oct 06 11:34:57 crc kubenswrapper[4989]: I1006 11:34:57.953609 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="066b101a-f170-4506-a60f-f82236f05912" path="/var/lib/kubelet/pods/066b101a-f170-4506-a60f-f82236f05912/volumes" Oct 06 11:34:58 crc kubenswrapper[4989]: I1006 11:34:58.937201 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:34:58 crc kubenswrapper[4989]: E1006 11:34:58.937963 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:35:03 crc kubenswrapper[4989]: I1006 11:35:03.196753 4989 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:35:03 crc kubenswrapper[4989]: I1006 11:35:03.273182 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:35:03 crc kubenswrapper[4989]: I1006 11:35:03.808859 4989 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4rm8x" podUID="7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" containerName="registry-server" containerID="cri-o://80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1" gracePeriod=2 Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.271277 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.420194 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5mrq\" (UniqueName: \"kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq\") pod \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.420270 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities\") pod \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.420721 4989 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content\") pod \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\" (UID: \"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc\") " Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.421198 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities" (OuterVolumeSpecName: "utilities") pod "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" (UID: "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.422776 4989 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.430986 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq" (OuterVolumeSpecName: "kube-api-access-m5mrq") pod "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" (UID: "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc"). InnerVolumeSpecName "kube-api-access-m5mrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.471638 4989 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" (UID: "7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.525119 4989 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5mrq\" (UniqueName: \"kubernetes.io/projected/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-kube-api-access-m5mrq\") on node \"crc\" DevicePath \"\"" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.525160 4989 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.821140 4989 generic.go:334] "Generic (PLEG): container finished" podID="7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" containerID="80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1" exitCode=0 Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.821230 4989 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rm8x" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.821220 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerDied","Data":"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1"} Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.821772 4989 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rm8x" event={"ID":"7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc","Type":"ContainerDied","Data":"40c36c92bdeb5eab1e581ecb4a61b27ab8d113d7a636f5227105799cb3c11d59"} Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.821806 4989 scope.go:117] "RemoveContainer" containerID="80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.851272 4989 scope.go:117] "RemoveContainer" containerID="56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52" Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.863421 4989 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:35:04 crc kubenswrapper[4989]: I1006 11:35:04.878567 4989 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4rm8x"] Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.203690 4989 scope.go:117] "RemoveContainer" containerID="6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.456301 4989 scope.go:117] "RemoveContainer" containerID="80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1" Oct 06 11:35:05 crc kubenswrapper[4989]: E1006 11:35:05.457360 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1\": container with ID starting with 80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1 not found: ID does not exist" containerID="80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.457411 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1"} err="failed to get container status \"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1\": rpc error: code = NotFound desc = could not find container \"80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1\": container with ID starting with 80f029ffc992e7400d5a413b0713317f4cbd81ec8cca86b3a5bf75891765e8a1 not found: ID does not exist" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.457443 4989 scope.go:117] "RemoveContainer" containerID="56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52" Oct 06 11:35:05 crc kubenswrapper[4989]: E1006 11:35:05.457919 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52\": container with ID starting with 56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52 not found: ID does not exist" containerID="56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.457949 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52"} err="failed to get container status \"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52\": rpc error: code = NotFound desc = could not find container \"56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52\": container with ID starting with 56ecd68b4949122712e2cedc4850e43be2256ab972ff63dfd60f3f2a3efa3b52 not found: ID does not exist" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.457967 4989 scope.go:117] "RemoveContainer" containerID="6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d" Oct 06 11:35:05 crc kubenswrapper[4989]: E1006 11:35:05.458239 4989 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d\": container with ID starting with 6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d not found: ID does not exist" containerID="6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.458265 4989 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d"} err="failed to get container status \"6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d\": rpc error: code = NotFound desc = could not find container \"6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d\": container with ID starting with 6434f1b1ba0f485a0b8cad1b24b1da8e788b75ed95b88df27c3d0bea3180403d not found: ID does not exist" Oct 06 11:35:05 crc kubenswrapper[4989]: I1006 11:35:05.947547 4989 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc" path="/var/lib/kubelet/pods/7f1ad003-62e8-4e8a-9e7a-8e57b1b445cc/volumes" Oct 06 11:35:11 crc kubenswrapper[4989]: I1006 11:35:11.936313 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:35:11 crc kubenswrapper[4989]: E1006 11:35:11.937249 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:35:24 crc kubenswrapper[4989]: I1006 11:35:24.936172 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:35:24 crc kubenswrapper[4989]: E1006 11:35:24.937015 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:35:36 crc kubenswrapper[4989]: I1006 11:35:36.937643 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:35:36 crc kubenswrapper[4989]: E1006 11:35:36.938556 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" Oct 06 11:35:50 crc kubenswrapper[4989]: I1006 11:35:50.937073 4989 scope.go:117] "RemoveContainer" containerID="5a938d345a25908e57ef3c08af5e848183d5d945c530c4c33c4cd0bd71d1fa36" Oct 06 11:35:50 crc kubenswrapper[4989]: E1006 11:35:50.938381 4989 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-v6vsq_openshift-machine-config-operator(f78ef9f2-a706-4a6c-8a9f-b232f97e9211)\"" pod="openshift-machine-config-operator/machine-config-daemon-v6vsq" podUID="f78ef9f2-a706-4a6c-8a9f-b232f97e9211" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070724641024453 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070724642017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070677506016522 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070677506015472 5ustar corecore